var/home/core/zuul-output/0000755000175000017500000000000015114065226014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114112741015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006626046115114112732017705 0ustar rootrootDec 03 16:55:43 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 16:55:43 crc restorecon[4757]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:43 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:55:44 crc restorecon[4757]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 16:55:44 crc kubenswrapper[4758]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:55:44 crc kubenswrapper[4758]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 16:55:44 crc kubenswrapper[4758]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:55:44 crc kubenswrapper[4758]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:55:44 crc kubenswrapper[4758]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 16:55:44 crc kubenswrapper[4758]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.911756 4758 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916581 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916654 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916668 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916756 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916771 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916782 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916792 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916808 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916826 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916838 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916850 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916862 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916873 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916885 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916898 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916909 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916919 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916931 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916941 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916952 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916962 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916976 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916986 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.916996 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917008 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917021 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917034 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917046 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917059 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917077 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917092 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917103 4758 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917120 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917134 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917146 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917158 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917168 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917177 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917185 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917194 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917202 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917209 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917218 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917225 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917233 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917243 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917251 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917258 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917266 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917274 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917282 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917290 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917297 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917306 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917314 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917324 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917333 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917340 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917348 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917355 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917363 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917370 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917378 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917385 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917392 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917401 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917408 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917416 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917426 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917436 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.917446 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918115 4758 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918145 4758 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918164 4758 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918179 4758 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918196 4758 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918210 4758 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918227 4758 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918243 4758 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918254 4758 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918266 4758 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918277 4758 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918287 4758 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918296 4758 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918306 4758 flags.go:64] FLAG: --cgroup-root="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918315 4758 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918325 4758 flags.go:64] FLAG: --client-ca-file="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918335 4758 flags.go:64] FLAG: --cloud-config="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918345 4758 flags.go:64] FLAG: --cloud-provider="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918353 4758 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918365 4758 flags.go:64] FLAG: --cluster-domain="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918375 4758 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918384 4758 flags.go:64] FLAG: --config-dir="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918393 4758 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918403 4758 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918415 4758 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918424 4758 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918434 4758 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918444 4758 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918453 4758 flags.go:64] FLAG: --contention-profiling="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918462 4758 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918471 4758 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918481 4758 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918490 4758 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918511 4758 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918521 4758 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918530 4758 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918539 4758 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918549 4758 flags.go:64] FLAG: --enable-server="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918558 4758 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918578 4758 flags.go:64] FLAG: --event-burst="100" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918590 4758 flags.go:64] FLAG: --event-qps="50" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918601 4758 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918613 4758 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918625 4758 flags.go:64] FLAG: --eviction-hard="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918661 4758 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918672 4758 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918736 4758 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918754 4758 flags.go:64] FLAG: --eviction-soft="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918765 4758 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918778 4758 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918791 4758 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918804 4758 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918817 4758 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918828 4758 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918840 4758 flags.go:64] FLAG: --feature-gates="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918855 4758 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918867 4758 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918879 4758 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918891 4758 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918903 4758 flags.go:64] FLAG: --healthz-port="10248" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918916 4758 flags.go:64] FLAG: --help="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918927 4758 flags.go:64] FLAG: --hostname-override="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918939 4758 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918951 4758 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918963 4758 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918974 4758 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918987 4758 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.918999 4758 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919010 4758 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919021 4758 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919032 4758 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919044 4758 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919056 4758 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919069 4758 flags.go:64] FLAG: --kube-reserved="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919082 4758 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919093 4758 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919106 4758 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919116 4758 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919128 4758 flags.go:64] FLAG: --lock-file="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919139 4758 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919150 4758 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919163 4758 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919183 4758 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919198 4758 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919210 4758 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919222 4758 flags.go:64] FLAG: --logging-format="text" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919234 4758 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919248 4758 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919260 4758 flags.go:64] FLAG: --manifest-url="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919273 4758 flags.go:64] FLAG: --manifest-url-header="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919291 4758 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919302 4758 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919317 4758 flags.go:64] FLAG: --max-pods="110" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919329 4758 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919343 4758 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919355 4758 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919366 4758 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919378 4758 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919390 4758 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919402 4758 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919432 4758 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919443 4758 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919455 4758 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919466 4758 flags.go:64] FLAG: --pod-cidr="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919478 4758 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919495 4758 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919507 4758 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919518 4758 flags.go:64] FLAG: --pods-per-core="0" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919529 4758 flags.go:64] FLAG: --port="10250" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919542 4758 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919553 4758 flags.go:64] FLAG: --provider-id="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919564 4758 flags.go:64] FLAG: --qos-reserved="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919576 4758 flags.go:64] FLAG: --read-only-port="10255" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919586 4758 flags.go:64] FLAG: --register-node="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919604 4758 flags.go:64] FLAG: --register-schedulable="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919612 4758 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919629 4758 flags.go:64] FLAG: --registry-burst="10" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919638 4758 flags.go:64] FLAG: --registry-qps="5" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919647 4758 flags.go:64] FLAG: --reserved-cpus="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919659 4758 flags.go:64] FLAG: --reserved-memory="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919671 4758 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919713 4758 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919723 4758 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919732 4758 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919741 4758 flags.go:64] FLAG: --runonce="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919751 4758 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919760 4758 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919771 4758 flags.go:64] FLAG: --seccomp-default="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919780 4758 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919789 4758 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919800 4758 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919810 4758 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919820 4758 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919830 4758 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919839 4758 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919848 4758 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919859 4758 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919872 4758 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919883 4758 flags.go:64] FLAG: --system-cgroups="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919894 4758 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919912 4758 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919921 4758 flags.go:64] FLAG: --tls-cert-file="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919930 4758 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919941 4758 flags.go:64] FLAG: --tls-min-version="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919949 4758 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919959 4758 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919971 4758 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919980 4758 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.919990 4758 flags.go:64] FLAG: --v="2" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.920002 4758 flags.go:64] FLAG: --version="false" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.920014 4758 flags.go:64] FLAG: --vmodule="" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.920025 4758 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.920035 4758 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920303 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920323 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920339 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920353 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920368 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920381 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920392 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920403 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920414 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920426 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920436 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920448 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920459 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920470 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920480 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920489 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920499 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920509 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920519 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920529 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920539 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920549 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920559 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920568 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920582 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920607 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920618 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920629 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920641 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920652 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920663 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920674 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920748 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920759 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920769 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920781 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920794 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920804 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920816 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920826 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920837 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920847 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920856 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920866 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920876 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920886 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920895 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920905 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920915 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920925 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920936 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920946 4758 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920957 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920968 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920979 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920987 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.920995 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921019 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921029 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921038 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921049 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921059 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921068 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921077 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921087 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921097 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921107 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921117 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921132 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921142 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.921151 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.921170 4758 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.932470 4758 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.932508 4758 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932639 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932651 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932661 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932736 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932759 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932770 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932780 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932790 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932802 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932811 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932819 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932828 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932837 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932845 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932854 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932862 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932869 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932877 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932886 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932894 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932901 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932909 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932917 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932925 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932932 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932940 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932948 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932955 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932962 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932975 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932983 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.932992 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933002 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933014 4758 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933025 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933033 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933042 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933051 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933059 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933067 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933077 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933085 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933093 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933101 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933109 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933119 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933129 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933137 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933145 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933153 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933161 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933168 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933176 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933184 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933191 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933199 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933206 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933214 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933222 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933230 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933240 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933251 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933260 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933268 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933277 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933288 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933298 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933308 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933317 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933326 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933335 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.933350 4758 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933581 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933594 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933603 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933612 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933622 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933630 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933638 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933646 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933654 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933662 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933670 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933677 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933716 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933726 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933736 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933744 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933752 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933759 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933770 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933780 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933788 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933797 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933807 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933817 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933826 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933834 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933842 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933849 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933857 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933866 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933874 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933882 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933889 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933897 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933905 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933913 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933921 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933928 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933937 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933944 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933952 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933960 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933968 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933976 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933983 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933991 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.933999 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934006 4758 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934017 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934026 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934035 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934043 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934051 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934059 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934066 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934077 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934085 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934093 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934101 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934109 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934117 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934125 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934132 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934140 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934148 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934158 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934169 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934177 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934185 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934193 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:55:44 crc kubenswrapper[4758]: W1203 16:55:44.934201 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.934213 4758 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.934447 4758 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.939453 4758 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.939580 4758 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.940480 4758 server.go:997] "Starting client certificate rotation" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.940519 4758 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.940929 4758 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-19 01:50:29.898538602 +0000 UTC Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.941076 4758 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 368h54m44.95747018s for next certificate rotation Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.953562 4758 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.956782 4758 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.971028 4758 log.go:25] "Validated CRI v1 runtime API" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.988258 4758 log.go:25] "Validated CRI v1 image API" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.991081 4758 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.997155 4758 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-16-50-27-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 16:55:44 crc kubenswrapper[4758]: I1203 16:55:44.997237 4758 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.027717 4758 manager.go:217] Machine: {Timestamp:2025-12-03 16:55:45.025699874 +0000 UTC m=+0.227076755 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8cc52a80-b3b7-4ab3-8e60-5535753f3e90 BootID:8c04589e-31a6-4e69-8c1b-e1795a06a3f9 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:58:91:a5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:58:91:a5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:62:0d:c9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3b:f6:44 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4d:0d:2a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c3:f0:5d Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:b4:0c:4f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:23:25:b1:8b:ff Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7e:32:48:25:2a:64 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.028190 4758 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.028437 4758 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.029167 4758 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.029499 4758 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.029565 4758 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.029997 4758 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.030020 4758 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.030429 4758 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.030496 4758 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.030856 4758 state_mem.go:36] "Initialized new in-memory state store" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.031010 4758 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.032156 4758 kubelet.go:418] "Attempting to sync node with API server" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.032206 4758 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.032277 4758 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.032308 4758 kubelet.go:324] "Adding apiserver pod source" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.032332 4758 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.035412 4758 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.035424 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.035862 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.036068 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.035945 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.036474 4758 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.037829 4758 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.038828 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.038997 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039098 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039213 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039326 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039417 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039501 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039609 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039738 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039830 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.039927 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.040021 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.040400 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.041774 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.042018 4758 server.go:1280] "Started kubelet" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.042567 4758 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.043082 4758 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.045414 4758 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 16:55:45 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.048335 4758 server.go:460] "Adding debug handlers to kubelet server" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.050565 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.050779 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 12:08:53.134149019 +0000 UTC Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.051428 4758 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.054054 4758 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.054093 4758 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.054348 4758 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.055604 4758 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.056114 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.056265 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.056449 4758 factory.go:55] Registering systemd factory Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.056481 4758 factory.go:221] Registration of the systemd container factory successfully Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.056455 4758 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dc2eb579778b3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:55:45.041750195 +0000 UTC m=+0.243127086,LastTimestamp:2025-12-03 16:55:45.041750195 +0000 UTC m=+0.243127086,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.057217 4758 factory.go:153] Registering CRI-O factory Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.057270 4758 factory.go:221] Registration of the crio container factory successfully Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.057331 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="200ms" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.057453 4758 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.057506 4758 factory.go:103] Registering Raw factory Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.057595 4758 manager.go:1196] Started watching for new ooms in manager Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.059217 4758 manager.go:319] Starting recovery of all containers Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073761 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073844 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073867 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073890 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073919 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073941 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073961 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.073989 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074011 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074031 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074050 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074068 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074152 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074183 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074213 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074231 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074252 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074287 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074305 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074324 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074341 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074360 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074391 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074419 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074452 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074473 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074504 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074524 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074547 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074570 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074589 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074610 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074671 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074734 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074763 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074782 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074802 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074822 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074843 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074914 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074967 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.074985 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075004 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075023 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075042 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075059 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075077 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075096 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075176 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075195 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075215 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075233 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075354 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075375 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075405 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075433 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075499 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075545 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075642 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075659 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075697 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075716 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075733 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075751 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075839 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075876 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075911 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075936 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.075964 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076006 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076041 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076065 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076201 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076227 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076249 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076273 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076297 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076322 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076346 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076382 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076475 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076499 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076523 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076775 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076815 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076841 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076877 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076905 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.076979 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077073 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077109 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077135 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077159 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077185 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077210 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077234 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077385 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077410 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077433 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077459 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077497 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077652 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077774 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.077809 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078088 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078197 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078271 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078302 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078339 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078378 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078413 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078449 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078560 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078584 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078650 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078675 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078731 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078754 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078779 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.078920 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079096 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079160 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079184 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079209 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079233 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079257 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079280 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079303 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079482 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079506 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079541 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079573 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079598 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079621 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079744 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079770 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079866 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079891 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079914 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079949 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079973 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.079997 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080021 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080044 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080148 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080181 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080204 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080227 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080343 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080374 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080409 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080434 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080527 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080608 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080632 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080742 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080777 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080844 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080869 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080894 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080948 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080966 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.080984 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081004 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081021 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081039 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081056 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081073 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081151 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081174 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081191 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081208 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081227 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081246 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081264 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.081284 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082569 4758 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082612 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082635 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082656 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082675 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082720 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082739 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082757 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.082816 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.083876 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.083944 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084200 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084244 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084303 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084334 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084362 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084417 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084446 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084494 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084528 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084556 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084642 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084712 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084785 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084816 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084844 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084884 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084909 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084936 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.084983 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.085022 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.085072 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.085096 4758 reconstruct.go:97] "Volume reconstruction finished" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.085113 4758 reconciler.go:26] "Reconciler: start to sync state" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.087744 4758 manager.go:324] Recovery completed Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.107269 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.109051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.109096 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.109120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.110113 4758 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.110136 4758 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.110172 4758 state_mem.go:36] "Initialized new in-memory state store" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.110380 4758 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.112934 4758 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.112990 4758 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.113056 4758 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.113243 4758 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.114832 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.114959 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.120235 4758 policy_none.go:49] "None policy: Start" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.123936 4758 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.123982 4758 state_mem.go:35] "Initializing new in-memory state store" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.156118 4758 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.181124 4758 manager.go:334] "Starting Device Plugin manager" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.181316 4758 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.181336 4758 server.go:79] "Starting device plugin registration server" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.181904 4758 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.181942 4758 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.182400 4758 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.182520 4758 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.182537 4758 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.191549 4758 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.214011 4758 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.214134 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.215308 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.215352 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.215364 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.215506 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.215967 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216075 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216421 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216495 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216622 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216862 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.216930 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217440 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217475 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217494 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217632 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217966 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.217947 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218079 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218370 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218403 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218415 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218573 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218729 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.218787 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219564 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219600 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219603 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219866 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.219889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.220111 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.220180 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.221008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.221043 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.221056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.258662 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="400ms" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.282850 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.285317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.285381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.285396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.285441 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.286153 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287282 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287332 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287361 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287384 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287408 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287456 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287487 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287627 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287669 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287726 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287764 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287797 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287841 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.287873 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389533 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389633 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389675 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389743 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389780 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389816 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389892 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389915 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389956 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389967 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390085 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389858 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390062 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.389941 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390068 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390209 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390247 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390283 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390310 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390315 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390378 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390368 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390328 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390517 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390552 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390603 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.390715 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.487289 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.489244 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.489287 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.489306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.489340 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.490113 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.539799 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.556156 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.572410 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-bafc1a523bef0ef11675c8dc294eca362cd380152741143cb6133a5eaf0485bc WatchSource:0}: Error finding container bafc1a523bef0ef11675c8dc294eca362cd380152741143cb6133a5eaf0485bc: Status 404 returned error can't find the container with id bafc1a523bef0ef11675c8dc294eca362cd380152741143cb6133a5eaf0485bc Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.579727 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.587028 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-14d10cec29b2b580b316dc879235402bc2480fb3efffa2e8a637a895fc8fc532 WatchSource:0}: Error finding container 14d10cec29b2b580b316dc879235402bc2480fb3efffa2e8a637a895fc8fc532: Status 404 returned error can't find the container with id 14d10cec29b2b580b316dc879235402bc2480fb3efffa2e8a637a895fc8fc532 Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.597831 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.607976 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.635121 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-670ad418b0e165b06912b57364fd4180450b8822039a592a0e36164f64d6d725 WatchSource:0}: Error finding container 670ad418b0e165b06912b57364fd4180450b8822039a592a0e36164f64d6d725: Status 404 returned error can't find the container with id 670ad418b0e165b06912b57364fd4180450b8822039a592a0e36164f64d6d725 Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.636944 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b7a91661e572653b20995ce5247d88ea78f9c1046618b38e1a6e31223e3e1a74 WatchSource:0}: Error finding container b7a91661e572653b20995ce5247d88ea78f9c1046618b38e1a6e31223e3e1a74: Status 404 returned error can't find the container with id b7a91661e572653b20995ce5247d88ea78f9c1046618b38e1a6e31223e3e1a74 Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.660254 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="800ms" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.860833 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.860949 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.890275 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.892100 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.892165 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.892180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:45 crc kubenswrapper[4758]: I1203 16:55:45.892216 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.892804 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 03 16:55:45 crc kubenswrapper[4758]: W1203 16:55:45.921296 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:45 crc kubenswrapper[4758]: E1203 16:55:45.921407 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.051138 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 19:21:20.438572913 +0000 UTC Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.052803 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 314h25m34.385782537s for next certificate rotation Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.051207 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.120992 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.121182 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b7a91661e572653b20995ce5247d88ea78f9c1046618b38e1a6e31223e3e1a74"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.122437 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c" exitCode=0 Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.122544 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.122613 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"670ad418b0e165b06912b57364fd4180450b8822039a592a0e36164f64d6d725"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.122784 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.124318 4758 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86" exitCode=0 Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.124367 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.124383 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"14d10cec29b2b580b316dc879235402bc2480fb3efffa2e8a637a895fc8fc532"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.124483 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.125130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.125167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.125177 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.125899 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.125938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.125948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.126090 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ffed3a292f0b121961ed8dee1c22d1b70ca5fd6594c46f218a21027f157a403e" exitCode=0 Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.126149 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ffed3a292f0b121961ed8dee1c22d1b70ca5fd6594c46f218a21027f157a403e"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.126209 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bafc1a523bef0ef11675c8dc294eca362cd380152741143cb6133a5eaf0485bc"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.126315 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127206 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127228 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127242 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127665 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127751 4758 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36" exitCode=0 Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127782 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127801 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"45e26395c24f9e7f3db65aa613e17ac2d51f652fe0d9ce7dbe892a21df0ea9f0"} Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.127872 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.128431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.128487 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.128501 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.128573 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.128605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.128621 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:46 crc kubenswrapper[4758]: W1203 16:55:46.224277 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:46 crc kubenswrapper[4758]: E1203 16:55:46.224740 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:46 crc kubenswrapper[4758]: W1203 16:55:46.251842 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 03 16:55:46 crc kubenswrapper[4758]: E1203 16:55:46.252023 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:55:46 crc kubenswrapper[4758]: E1203 16:55:46.461454 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="1.6s" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.694582 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.698921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.698983 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.698994 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:46 crc kubenswrapper[4758]: I1203 16:55:46.699033 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.134455 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.134515 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.134529 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.134659 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.136191 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.136215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.136224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.137855 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.137876 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.137889 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.137943 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.138936 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.138961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.138972 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.141731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.141748 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.141759 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.141771 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.143598 4758 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062" exitCode=0 Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.143629 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.143796 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.144600 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.144621 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.144633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.146077 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"567a6727629cd00ae7f791b31551abbc9fad1a640cff7e997ced687bb5e8bca7"} Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.146153 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.147090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.147108 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:47 crc kubenswrapper[4758]: I1203 16:55:47.147116 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.154944 4758 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b" exitCode=0 Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.155588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b"} Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.155887 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.157430 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.157488 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.157514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.162625 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.162675 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238"} Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.163311 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.163860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.163917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.163945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.165093 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.165151 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:48 crc kubenswrapper[4758]: I1203 16:55:48.165176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.170796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7"} Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.170875 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0"} Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.170897 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811"} Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.170915 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.170977 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.171937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.171987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:49 crc kubenswrapper[4758]: I1203 16:55:49.172004 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.178369 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426"} Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.178418 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3"} Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.178594 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.179945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.180004 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.180024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.678416 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.678757 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.680532 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.680593 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.680616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.970825 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.971040 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.971094 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.972894 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.972975 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:50 crc kubenswrapper[4758]: I1203 16:55:50.972992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:51 crc kubenswrapper[4758]: I1203 16:55:51.180552 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:51 crc kubenswrapper[4758]: I1203 16:55:51.181537 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:51 crc kubenswrapper[4758]: I1203 16:55:51.181568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:51 crc kubenswrapper[4758]: I1203 16:55:51.181577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:52 crc kubenswrapper[4758]: I1203 16:55:52.234118 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:52 crc kubenswrapper[4758]: I1203 16:55:52.234351 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:55:52 crc kubenswrapper[4758]: I1203 16:55:52.234402 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:52 crc kubenswrapper[4758]: I1203 16:55:52.235802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:52 crc kubenswrapper[4758]: I1203 16:55:52.235860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:52 crc kubenswrapper[4758]: I1203 16:55:52.235877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:53 crc kubenswrapper[4758]: I1203 16:55:53.720937 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:53 crc kubenswrapper[4758]: I1203 16:55:53.721168 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:53 crc kubenswrapper[4758]: I1203 16:55:53.722640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:53 crc kubenswrapper[4758]: I1203 16:55:53.722713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:53 crc kubenswrapper[4758]: I1203 16:55:53.722725 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.894350 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.894738 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.896575 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.896631 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.896644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.976387 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.976706 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.978130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.978191 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:54 crc kubenswrapper[4758]: I1203 16:55:54.978204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:55 crc kubenswrapper[4758]: E1203 16:55:55.191719 4758 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 16:55:55 crc kubenswrapper[4758]: I1203 16:55:55.311183 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:55:55 crc kubenswrapper[4758]: I1203 16:55:55.311451 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:55 crc kubenswrapper[4758]: I1203 16:55:55.313128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:55 crc kubenswrapper[4758]: I1203 16:55:55.313192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:55 crc kubenswrapper[4758]: I1203 16:55:55.313207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.125622 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.125931 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.127783 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.127827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.127837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.130804 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.193840 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.195446 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.195478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.195488 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.198903 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.519558 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.519909 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.521620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.521664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.521738 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:56 crc kubenswrapper[4758]: E1203 16:55:56.700567 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 03 16:55:56 crc kubenswrapper[4758]: I1203 16:55:56.933497 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.071174 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.195955 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.196784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.196819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.196830 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:57 crc kubenswrapper[4758]: W1203 16:55:57.552663 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.552775 4758 trace.go:236] Trace[95318495]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:55:47.551) (total time: 10001ms): Dec 03 16:55:57 crc kubenswrapper[4758]: Trace[95318495]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:55:57.552) Dec 03 16:55:57 crc kubenswrapper[4758]: Trace[95318495]: [10.001467068s] [10.001467068s] END Dec 03 16:55:57 crc kubenswrapper[4758]: E1203 16:55:57.552797 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.576277 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.576391 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.600047 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 16:55:57 crc kubenswrapper[4758]: I1203 16:55:57.600132 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.198663 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.199756 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.199784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.199793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.301279 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.302776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.302822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.302832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:55:58 crc kubenswrapper[4758]: I1203 16:55:58.302864 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:55:59 crc kubenswrapper[4758]: I1203 16:55:59.934713 4758 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 16:55:59 crc kubenswrapper[4758]: I1203 16:55:59.934789 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 16:56:01 crc kubenswrapper[4758]: I1203 16:56:01.511003 4758 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.240138 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.240356 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.241548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.241592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.241607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.243993 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:56:02 crc kubenswrapper[4758]: E1203 16:56:02.579912 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.582843 4758 trace.go:236] Trace[1492657652]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:55:48.901) (total time: 13681ms): Dec 03 16:56:02 crc kubenswrapper[4758]: Trace[1492657652]: ---"Objects listed" error: 13681ms (16:56:02.582) Dec 03 16:56:02 crc kubenswrapper[4758]: Trace[1492657652]: [13.681194895s] [13.681194895s] END Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.582875 4758 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.583347 4758 trace.go:236] Trace[171786253]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:55:47.926) (total time: 14656ms): Dec 03 16:56:02 crc kubenswrapper[4758]: Trace[171786253]: ---"Objects listed" error: 14656ms (16:56:02.583) Dec 03 16:56:02 crc kubenswrapper[4758]: Trace[171786253]: [14.65696243s] [14.65696243s] END Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.583368 4758 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.583542 4758 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.584724 4758 trace.go:236] Trace[1128419527]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:55:48.832) (total time: 13751ms): Dec 03 16:56:02 crc kubenswrapper[4758]: Trace[1128419527]: ---"Objects listed" error: 13751ms (16:56:02.584) Dec 03 16:56:02 crc kubenswrapper[4758]: Trace[1128419527]: [13.751710081s] [13.751710081s] END Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.584757 4758 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.783326 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38242->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.783353 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38244->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.783401 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38242->192.168.126.11:17697: read: connection reset by peer" Dec 03 16:56:02 crc kubenswrapper[4758]: I1203 16:56:02.783423 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38244->192.168.126.11:17697: read: connection reset by peer" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.075490 4758 apiserver.go:52] "Watching apiserver" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.079054 4758 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.079418 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.079907 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.080087 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.080214 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.080216 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.080109 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.080646 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.080715 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.080770 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.080820 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.082239 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.082560 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.082893 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.086606 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.086656 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.086666 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.086724 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.087108 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.087194 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.148507 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.155795 4758 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.161089 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.170515 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.178369 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188053 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188096 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188122 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188143 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188164 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188182 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188197 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188213 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188228 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188242 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188257 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188272 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188289 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188305 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188323 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188340 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188354 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188376 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188447 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188474 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188494 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188510 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188528 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188544 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188579 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188596 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188615 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188634 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188651 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188668 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188662 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188727 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188706 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188814 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188843 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188867 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188897 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188923 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188954 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188986 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189046 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189073 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189096 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189121 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189144 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189166 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189189 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189218 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189240 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189261 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189284 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189307 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189330 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189352 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189372 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189395 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189423 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189447 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189474 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189498 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189522 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189545 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189572 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189598 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189624 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189648 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189669 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189710 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189732 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189770 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189792 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189817 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189837 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189858 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189879 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189906 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189950 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189973 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189996 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190019 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190042 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190065 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190086 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190111 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190133 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190152 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190172 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190195 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190221 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190245 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190294 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190270 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190318 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190343 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190365 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190386 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190407 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190450 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190469 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190492 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190513 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190537 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190589 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190613 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190638 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190733 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190755 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190804 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190828 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190852 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190875 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190897 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190918 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190941 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190971 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190996 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191018 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191045 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191069 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191094 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191120 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191144 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191166 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191187 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191230 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191279 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191303 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191326 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191374 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191401 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191424 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191560 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191596 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191622 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191645 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191669 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191821 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191850 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191880 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191906 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191932 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191958 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191983 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192006 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192038 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192060 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192083 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192108 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192132 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192155 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192185 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192210 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192236 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192260 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192281 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192316 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192340 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192365 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192392 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192415 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188864 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192447 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188874 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188923 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.188943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189043 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189095 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189137 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189211 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192696 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192781 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192801 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189383 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189425 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189480 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189501 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189501 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189567 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189735 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190544 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190580 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190782 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190792 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.190933 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191033 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191090 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191242 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191242 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191327 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191519 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191576 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191787 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191782 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191818 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191814 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.191853 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192217 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192419 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192530 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192884 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193111 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193191 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193272 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193279 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193464 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193597 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193630 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193633 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193760 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.193812 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194111 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194137 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194167 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194216 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194242 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194521 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194554 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194566 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194588 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194850 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.194936 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.195076 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.195392 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.195442 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.195881 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.195931 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196166 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196282 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196337 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196445 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196541 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196636 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196764 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196834 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196994 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.196995 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.197140 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.197366 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.197504 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.198438 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.198574 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.198633 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.198943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199209 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199340 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199292 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199239 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199211 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199550 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.199723 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.200042 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.200121 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.200532 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.200582 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.200650 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.200816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201131 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201189 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201300 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201374 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201581 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.201774 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:56:03.701752956 +0000 UTC m=+18.903129817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201770 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.201891 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.202189 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.202912 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.203008 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.203180 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.204342 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.204831 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205087 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205161 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205477 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205487 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205555 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205749 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.192478 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206197 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.189299 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206239 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206278 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206302 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206320 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206343 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206366 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206391 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206407 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206426 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206446 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206464 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206480 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206495 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206511 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206526 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206542 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206557 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206576 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206592 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206608 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206624 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206641 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206658 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206742 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206770 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206788 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206849 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206871 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206892 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206937 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206957 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206976 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207002 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207018 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207051 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207070 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207091 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207111 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207129 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207220 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207234 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207245 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207257 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207267 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207284 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207294 4758 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207303 4758 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207314 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207325 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207334 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207347 4758 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207356 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207371 4758 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207382 4758 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207392 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207400 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207409 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207420 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207430 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207439 4758 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207449 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207459 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207498 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207508 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207517 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207528 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207538 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207549 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207559 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207569 4758 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207578 4758 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207586 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207595 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207603 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207611 4758 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207619 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207628 4758 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207637 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207646 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207656 4758 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207665 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207673 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207699 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207710 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207722 4758 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207733 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207742 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207752 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207762 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207771 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207780 4758 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207790 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207799 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207809 4758 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207817 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207777 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207825 4758 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208033 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208066 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208081 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208096 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208116 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208131 4758 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208145 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208159 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208173 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208187 4758 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208201 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208219 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208406 4758 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208430 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208446 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208458 4758 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208472 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208487 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208503 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208516 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208772 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208823 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208845 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208862 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208875 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208885 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208897 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208909 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208919 4758 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208930 4758 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208941 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208951 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208962 4758 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208973 4758 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208982 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208994 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209003 4758 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209013 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209022 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209031 4758 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209041 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209051 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209060 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209069 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209088 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209099 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209109 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209119 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209128 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209137 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209147 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209157 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209167 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209176 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209186 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209196 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209207 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209216 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209226 4758 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209235 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209246 4758 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209255 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209266 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209275 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209284 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209294 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209303 4758 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209313 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208792 4758 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.212917 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206421 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206564 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.206729 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207049 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207108 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207152 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207203 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207212 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207858 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.207995 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208086 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208102 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.208385 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213536 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.213575 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:03.713550064 +0000 UTC m=+18.914926925 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213607 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213698 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208388 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208426 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213902 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213999 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214096 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214122 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208516 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.208634 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.209397 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.214232 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:03.714214011 +0000 UTC m=+18.915590872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.209914 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.210065 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.210424 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.210859 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.211145 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.211541 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.212801 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.212855 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.212673 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.212909 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213060 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.213176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.205092 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214558 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.215156 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214613 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214628 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214659 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214852 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.214898 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.215071 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.215148 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.215868 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.215950 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.215996 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.216346 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.216357 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.216403 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.216885 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.217944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.218129 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.218652 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.219976 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.220899 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.220931 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.221212 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.221594 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.223744 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.224809 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.225742 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.226316 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.226988 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.227692 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.228840 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.228869 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.228885 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.228952 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:03.728934885 +0000 UTC m=+18.930311746 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.229190 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.229987 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.230013 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.230030 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.230099 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:03.730076895 +0000 UTC m=+18.931453856 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.230718 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238" exitCode=255 Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.230763 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.232860 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.236771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.237804 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.238835 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.247750 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.249865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.251848 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.253041 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.252003 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.255403 4758 scope.go:117] "RemoveContainer" containerID="4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.264336 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.280496 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.294265 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.307979 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310195 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310227 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310237 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310247 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310255 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310287 4758 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310296 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310307 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310315 4758 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310323 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310331 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310339 4758 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310347 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310355 4758 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310362 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310373 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310381 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310389 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310397 4758 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310405 4758 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310413 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310421 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310429 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310437 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310445 4758 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310453 4758 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310461 4758 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310469 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310478 4758 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310485 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310493 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310501 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310510 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310518 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310526 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310533 4758 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310544 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310552 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310561 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310569 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310577 4758 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310587 4758 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310596 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310604 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310611 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310620 4758 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310627 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310636 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310643 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310652 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310660 4758 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310669 4758 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310692 4758 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310700 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310708 4758 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310716 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310725 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310733 4758 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310741 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310749 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310757 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310766 4758 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310774 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310783 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310792 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310800 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310808 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310816 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310814 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310826 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310870 4758 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310882 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.310961 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.315963 4758 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.316054 4758 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.324057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.324102 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.324113 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.324129 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.324142 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.337999 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.342976 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.348928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.348961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.348970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.348984 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.348994 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.354774 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.370260 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.380834 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.384413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.384447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.384457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.384471 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.384481 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.399040 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.406428 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.420110 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.434881 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.460179 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.506835 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.506877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.506888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.506905 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.506915 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.536066 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.543843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.543879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.543889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.543905 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.543916 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.573608 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.573783 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.580029 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.580076 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.580088 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.580105 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.580116 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.682842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.682892 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.682907 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.682928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.682942 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.717239 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.717338 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.717396 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.717538 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.717598 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:04.717581448 +0000 UTC m=+19.918958319 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.717663 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:56:04.717655179 +0000 UTC m=+19.919032040 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.717755 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.717784 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:04.717775743 +0000 UTC m=+19.919152604 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.785929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.785996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.786009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.786026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.786038 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.818477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.818533 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818701 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818726 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818739 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818754 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818790 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818810 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818817 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:04.818787122 +0000 UTC m=+20.020163983 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: E1203 16:56:03.818883 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:04.818861654 +0000 UTC m=+20.020238595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.891994 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.892045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.892054 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.892071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.892080 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.994603 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.994653 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.994665 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.994700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:03 crc kubenswrapper[4758]: I1203 16:56:03.994715 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:03Z","lastTransitionTime":"2025-12-03T16:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.097012 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.097050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.097061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.097079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.097089 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.199985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.200026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.200039 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.200055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.200067 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.235290 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.235346 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.235360 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"335c019609069581f93b1b32af303684730e463fd411b262b53f0bbf75c98498"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.237928 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.237986 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6cf5d61261153470421a7ac87d5e27e81e6231cfc00a1ab437f0e8a580eacd65"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.239990 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.241656 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.241875 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.242599 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8a8fd766d3c5adcba188bc090a931a542c0c30e00186e214e2b6ad59234984c7"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.261743 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zjq67"] Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.262319 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-zwcmr"] Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.262481 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-6hk5h"] Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.262481 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.262601 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.262896 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.264052 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.264061 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.264810 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.265022 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-fdssg"] Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.265393 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.272776 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.272814 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.273071 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.273459 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.274570 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9p24"] Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.275284 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.275298 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.275390 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.275440 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.275489 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.275526 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.277286 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.277912 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.277939 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.278493 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.278491 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.278865 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.278982 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.279424 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.279573 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.291657 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.291665 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.302332 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.302376 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.302387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.302403 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.302414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.308604 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.330957 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-cni-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331004 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-os-release\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331029 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-socket-dir-parent\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331147 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-hostroot\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-daemon-config\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331194 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-k8s-cni-cncf-io\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331278 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-proxy-tls\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331309 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b5429d1-1fc3-4603-93ba-b57b33c2f585-cni-binary-copy\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331374 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0dc871d6-da79-41c8-b479-0a0d193d4c78-hosts-file\") pod \"node-resolver-6hk5h\" (UID: \"0dc871d6-da79-41c8-b479-0a0d193d4c78\") " pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331487 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-conf-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331550 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-multus-certs\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331578 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-cnibin\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331597 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-netns\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331620 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/06439031-4230-488c-9422-97448eeac39d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331759 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kjpf\" (UniqueName: \"kubernetes.io/projected/06439031-4230-488c-9422-97448eeac39d-kube-api-access-7kjpf\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331837 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-system-cni-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331865 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftht5\" (UniqueName: \"kubernetes.io/projected/0dc871d6-da79-41c8-b479-0a0d193d4c78-kube-api-access-ftht5\") pod \"node-resolver-6hk5h\" (UID: \"0dc871d6-da79-41c8-b479-0a0d193d4c78\") " pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331888 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-rootfs\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331932 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgt44\" (UniqueName: \"kubernetes.io/projected/4b5429d1-1fc3-4603-93ba-b57b33c2f585-kube-api-access-tgt44\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331956 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-os-release\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.331990 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-cni-multus\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332016 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-system-cni-dir\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332038 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-mcd-auth-proxy-config\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332087 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl84d\" (UniqueName: \"kubernetes.io/projected/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-kube-api-access-rl84d\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332110 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-cni-bin\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332131 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-etc-kubernetes\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332151 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-cnibin\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332173 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332198 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-kubelet\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.332219 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06439031-4230-488c-9422-97448eeac39d-cni-binary-copy\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.333410 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.371318 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.395882 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.404595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.404631 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.404638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.404652 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.404662 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.407294 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.418728 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.431019 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432611 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-conf-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-multus-certs\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-slash\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432690 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-node-log\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432705 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47xn2\" (UniqueName: \"kubernetes.io/projected/b40dc3c2-0bc6-447b-ae72-d71650d53f58-kube-api-access-47xn2\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432719 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-cnibin\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432734 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-netns\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432749 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/06439031-4230-488c-9422-97448eeac39d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432763 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kjpf\" (UniqueName: \"kubernetes.io/projected/06439031-4230-488c-9422-97448eeac39d-kube-api-access-7kjpf\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432778 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-system-cni-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432795 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftht5\" (UniqueName: \"kubernetes.io/projected/0dc871d6-da79-41c8-b479-0a0d193d4c78-kube-api-access-ftht5\") pod \"node-resolver-6hk5h\" (UID: \"0dc871d6-da79-41c8-b479-0a0d193d4c78\") " pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432811 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-kubelet\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432825 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-systemd\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432838 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-netd\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgt44\" (UniqueName: \"kubernetes.io/projected/4b5429d1-1fc3-4603-93ba-b57b33c2f585-kube-api-access-tgt44\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432872 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-os-release\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432885 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-rootfs\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432899 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-config\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432913 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-script-lib\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432928 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-cni-multus\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432943 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-system-cni-dir\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432957 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-mcd-auth-proxy-config\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432972 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-log-socket\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.432986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-bin\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433007 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-env-overrides\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433023 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-cni-bin\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433036 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-etc-kubernetes\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433050 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-cnibin\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433064 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433080 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl84d\" (UniqueName: \"kubernetes.io/projected/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-kube-api-access-rl84d\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433095 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-systemd-units\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433109 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-ovn\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433125 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-ovn-kubernetes\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433139 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-kubelet\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06439031-4230-488c-9422-97448eeac39d-cni-binary-copy\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433170 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-etc-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-netns\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-cni-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433224 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-os-release\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433239 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-socket-dir-parent\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433253 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-hostroot\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433280 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-daemon-config\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433294 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-var-lib-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433311 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-k8s-cni-cncf-io\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433327 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-proxy-tls\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433342 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433358 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433374 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovn-node-metrics-cert\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433397 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b5429d1-1fc3-4603-93ba-b57b33c2f585-cni-binary-copy\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433412 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0dc871d6-da79-41c8-b479-0a0d193d4c78-hosts-file\") pod \"node-resolver-6hk5h\" (UID: \"0dc871d6-da79-41c8-b479-0a0d193d4c78\") " pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433475 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0dc871d6-da79-41c8-b479-0a0d193d4c78-hosts-file\") pod \"node-resolver-6hk5h\" (UID: \"0dc871d6-da79-41c8-b479-0a0d193d4c78\") " pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433505 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-conf-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433524 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-multus-certs\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433560 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-cnibin\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.433582 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-netns\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434127 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/06439031-4230-488c-9422-97448eeac39d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434138 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434223 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-socket-dir-parent\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434224 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-system-cni-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434280 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-kubelet\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434403 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-hostroot\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434642 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-os-release\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434703 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-rootfs\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434736 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-cni-multus\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.434758 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-system-cni-dir\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435003 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-var-lib-cni-bin\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435036 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06439031-4230-488c-9422-97448eeac39d-cni-binary-copy\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435047 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-cni-dir\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435047 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06439031-4230-488c-9422-97448eeac39d-cnibin\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435089 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-host-run-k8s-cni-cncf-io\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435085 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-etc-kubernetes\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435141 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b5429d1-1fc3-4603-93ba-b57b33c2f585-os-release\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4b5429d1-1fc3-4603-93ba-b57b33c2f585-multus-daemon-config\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.435572 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b5429d1-1fc3-4603-93ba-b57b33c2f585-cni-binary-copy\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.436081 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-mcd-auth-proxy-config\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.440148 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-proxy-tls\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.449382 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.450171 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl84d\" (UniqueName: \"kubernetes.io/projected/2e7d0dba-00e7-4a9d-915e-e31788c7a2bb-kube-api-access-rl84d\") pod \"machine-config-daemon-fdssg\" (UID: \"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\") " pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.459421 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kjpf\" (UniqueName: \"kubernetes.io/projected/06439031-4230-488c-9422-97448eeac39d-kube-api-access-7kjpf\") pod \"multus-additional-cni-plugins-zjq67\" (UID: \"06439031-4230-488c-9422-97448eeac39d\") " pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.468428 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.471100 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftht5\" (UniqueName: \"kubernetes.io/projected/0dc871d6-da79-41c8-b479-0a0d193d4c78-kube-api-access-ftht5\") pod \"node-resolver-6hk5h\" (UID: \"0dc871d6-da79-41c8-b479-0a0d193d4c78\") " pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.471145 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgt44\" (UniqueName: \"kubernetes.io/projected/4b5429d1-1fc3-4603-93ba-b57b33c2f585-kube-api-access-tgt44\") pod \"multus-zwcmr\" (UID: \"4b5429d1-1fc3-4603-93ba-b57b33c2f585\") " pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.487919 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.497901 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.506787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.506814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.506844 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.506858 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.506867 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.513370 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.524697 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-log-socket\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534079 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-bin\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534109 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-env-overrides\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534132 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-systemd-units\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-ovn\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534182 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-ovn-kubernetes\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534208 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-etc-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534198 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-bin\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534275 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-systemd-units\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534276 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-netns\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534302 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-ovn-kubernetes\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534335 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-ovn\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534335 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-etc-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-netns\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534377 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-log-socket\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-var-lib-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534413 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534432 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534454 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovn-node-metrics-cert\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534460 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534507 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-slash\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534528 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-node-log\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534547 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47xn2\" (UniqueName: \"kubernetes.io/projected/b40dc3c2-0bc6-447b-ae72-d71650d53f58-kube-api-access-47xn2\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534573 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-kubelet\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534591 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-systemd\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-netd\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534623 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-env-overrides\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534635 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-config\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534654 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-script-lib\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534909 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-slash\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534435 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534950 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-systemd\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534971 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-kubelet\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.534933 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-node-log\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.535037 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-netd\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.535320 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-script-lib\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.535373 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-config\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.535384 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-var-lib-openvswitch\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.538488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovn-node-metrics-cert\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.542988 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.559229 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47xn2\" (UniqueName: \"kubernetes.io/projected/b40dc3c2-0bc6-447b-ae72-d71650d53f58-kube-api-access-47xn2\") pod \"ovnkube-node-p9p24\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.561223 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.574331 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zjq67" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.581108 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zwcmr" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.581224 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.589469 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6hk5h" Dec 03 16:56:04 crc kubenswrapper[4758]: W1203 16:56:04.592459 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06439031_4230_488c_9422_97448eeac39d.slice/crio-595d56518d918ab7df6900c672771b6bfdf27bb49440a728004e1fce90c8e25b WatchSource:0}: Error finding container 595d56518d918ab7df6900c672771b6bfdf27bb49440a728004e1fce90c8e25b: Status 404 returned error can't find the container with id 595d56518d918ab7df6900c672771b6bfdf27bb49440a728004e1fce90c8e25b Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.595767 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:56:04 crc kubenswrapper[4758]: W1203 16:56:04.599434 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b5429d1_1fc3_4603_93ba_b57b33c2f585.slice/crio-9347445fe9d01b91d7f69ac6c31848dd88326b2588cf048178f3886293e941d4 WatchSource:0}: Error finding container 9347445fe9d01b91d7f69ac6c31848dd88326b2588cf048178f3886293e941d4: Status 404 returned error can't find the container with id 9347445fe9d01b91d7f69ac6c31848dd88326b2588cf048178f3886293e941d4 Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.599536 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.601921 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:04 crc kubenswrapper[4758]: W1203 16:56:04.604953 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dc871d6_da79_41c8_b479_0a0d193d4c78.slice/crio-d287d4fec96bac335d98eac03bb717469722c52afc33c9ab8980b356b0fd21ca WatchSource:0}: Error finding container d287d4fec96bac335d98eac03bb717469722c52afc33c9ab8980b356b0fd21ca: Status 404 returned error can't find the container with id d287d4fec96bac335d98eac03bb717469722c52afc33c9ab8980b356b0fd21ca Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.611119 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.611152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.611162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.611176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.611185 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.617163 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:04 crc kubenswrapper[4758]: W1203 16:56:04.626502 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e7d0dba_00e7_4a9d_915e_e31788c7a2bb.slice/crio-8f3cf781ea185382b3b7fa6cbd562801508f1aeec723204b758969d3fbf03ded WatchSource:0}: Error finding container 8f3cf781ea185382b3b7fa6cbd562801508f1aeec723204b758969d3fbf03ded: Status 404 returned error can't find the container with id 8f3cf781ea185382b3b7fa6cbd562801508f1aeec723204b758969d3fbf03ded Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.718086 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.718124 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.718135 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.718152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.718164 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.739991 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.740114 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.740138 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.740226 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.740271 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:06.740257325 +0000 UTC m=+21.941634186 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.740319 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:56:06.740311546 +0000 UTC m=+21.941688397 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.740372 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.740391 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:06.740386318 +0000 UTC m=+21.941763179 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.822818 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.822844 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.822852 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.822869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.822878 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.840819 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.840859 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.840985 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841004 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841015 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841026 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841060 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841076 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841064 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:06.841049819 +0000 UTC m=+22.042426680 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:04 crc kubenswrapper[4758]: E1203 16:56:04.841138 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:06.841122191 +0000 UTC m=+22.042499052 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.925843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.925891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.925901 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.925918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:04 crc kubenswrapper[4758]: I1203 16:56:04.925929 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:04Z","lastTransitionTime":"2025-12-03T16:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.031392 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.031427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.031439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.031456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.031468 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.113947 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.113954 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:05 crc kubenswrapper[4758]: E1203 16:56:05.114068 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.114089 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:05 crc kubenswrapper[4758]: E1203 16:56:05.114167 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:05 crc kubenswrapper[4758]: E1203 16:56:05.114243 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.117370 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.118456 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.120147 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.121220 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.122907 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.123776 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.124901 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.128102 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.129351 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.131066 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.131803 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.133404 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.133546 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.133568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.133576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.133592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.133601 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.134447 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.135433 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.136772 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.137463 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.138843 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.139468 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.140353 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.141756 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.142671 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.143658 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.145268 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.145705 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.146960 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.149243 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.151019 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.153482 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.154589 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.155854 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.157372 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.158075 4758 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.158285 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.161075 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.161898 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.162718 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.165179 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.166304 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.167652 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.168834 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.170041 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.170647 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.171318 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.172560 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.173906 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.173911 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.174801 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.176049 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.176936 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.178349 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.179130 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.179896 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.180495 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.181783 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.182511 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.183570 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.198208 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.227314 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.235279 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.235313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.235324 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.235341 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.235352 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.247816 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3" exitCode=0 Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.247890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.247919 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"c3de506de982657a3e05ab139a6d51b0375cba0005e39dbaba3b902b37ea6dce"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.247894 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.250579 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.250616 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.250628 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"8f3cf781ea185382b3b7fa6cbd562801508f1aeec723204b758969d3fbf03ded"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.252613 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerStarted","Data":"3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.252638 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerStarted","Data":"9347445fe9d01b91d7f69ac6c31848dd88326b2588cf048178f3886293e941d4"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.255294 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerStarted","Data":"d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.255332 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerStarted","Data":"595d56518d918ab7df6900c672771b6bfdf27bb49440a728004e1fce90c8e25b"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.260736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6hk5h" event={"ID":"0dc871d6-da79-41c8-b479-0a0d193d4c78","Type":"ContainerStarted","Data":"493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.260785 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6hk5h" event={"ID":"0dc871d6-da79-41c8-b479-0a0d193d4c78","Type":"ContainerStarted","Data":"d287d4fec96bac335d98eac03bb717469722c52afc33c9ab8980b356b0fd21ca"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.269904 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.290654 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.305716 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.322139 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.333756 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.337837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.337897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.337927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.337984 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.338001 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.349002 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.393395 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.438144 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.440140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.440294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.440362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.440446 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.440518 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.472862 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.516961 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.537736 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.542239 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.542266 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.542273 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.542288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.542297 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.551279 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.568086 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.601957 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.623939 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.640638 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.645189 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.645369 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.645455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.645521 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.645573 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.661666 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.675695 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.691764 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.747514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.747543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.747554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.747568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.747579 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.849806 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.849838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.849846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.849859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.849867 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.953146 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.953184 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.953199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.953446 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:05 crc kubenswrapper[4758]: I1203 16:56:05.953461 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:05Z","lastTransitionTime":"2025-12-03T16:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.055349 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.055381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.055389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.055403 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.055414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.159935 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.160211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.160225 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.160241 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.160253 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.262639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.262691 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.262701 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.262717 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.262729 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.265607 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.265638 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.265647 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.265655 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.266982 4758 generic.go:334] "Generic (PLEG): container finished" podID="06439031-4230-488c-9422-97448eeac39d" containerID="d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7" exitCode=0 Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.267018 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerDied","Data":"d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.268260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.284126 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.302714 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.314723 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.338887 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.352092 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.365767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.365804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.365813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.365826 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.365838 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.366352 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.378556 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.393729 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.409761 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.423161 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.434281 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.446819 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.458979 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.468116 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.468142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.468151 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.468163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.468171 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.472748 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.495845 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.509810 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.520124 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.534090 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.542289 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.554823 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.555368 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.558979 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.568664 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.570577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.570619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.570633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.570647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.570969 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.580647 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.593326 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.605080 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.614591 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.625136 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.634278 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.656134 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.670283 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.672912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.672943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.672952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.672968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.672978 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.685648 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.698521 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.710491 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.725117 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.744082 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.758260 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.759542 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.759690 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:56:10.759653403 +0000 UTC m=+25.961030254 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.759992 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.760286 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.760128 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.760556 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:10.760540397 +0000 UTC m=+25.961917258 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.760367 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.760785 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:10.760776653 +0000 UTC m=+25.962153504 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.770980 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.775048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.775083 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.775091 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.775108 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.775120 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.784012 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.797916 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.861400 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.861445 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861554 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861570 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861580 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861620 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:10.861608338 +0000 UTC m=+26.062985199 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861653 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861730 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861744 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:06 crc kubenswrapper[4758]: E1203 16:56:06.861804 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:10.861783983 +0000 UTC m=+26.063160884 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.877021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.877057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.877065 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.877081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.877091 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.936803 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.941201 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.944472 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.949990 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.960787 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.971171 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.978939 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.978974 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.978987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.979002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.979012 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:06Z","lastTransitionTime":"2025-12-03T16:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.981316 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:06 crc kubenswrapper[4758]: I1203 16:56:06.993023 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.002472 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.024463 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.035393 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.052999 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.065385 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.076053 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.081918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.081953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.081961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.081978 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.081988 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.089000 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.111134 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.113345 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.113389 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:07 crc kubenswrapper[4758]: E1203 16:56:07.113452 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:07 crc kubenswrapper[4758]: E1203 16:56:07.113529 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.113628 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:07 crc kubenswrapper[4758]: E1203 16:56:07.113794 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.124326 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.136335 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.147811 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.184152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.184193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.184203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.184219 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.184230 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.200152 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.221988 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.233716 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.243211 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.256427 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.270945 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.274385 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.274425 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.275885 4758 generic.go:334] "Generic (PLEG): container finished" podID="06439031-4230-488c-9422-97448eeac39d" containerID="531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f" exitCode=0 Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.275998 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerDied","Data":"531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.289164 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.290580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.290610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.290619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.290633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.290642 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.306566 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.318227 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.331135 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.349494 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.362493 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.373996 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.393647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.393801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.393813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.393828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.393838 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.397220 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.412130 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.427055 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.438658 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.452993 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.466322 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.480266 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.497398 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.497429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.497440 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.497457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.497469 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.509488 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.547392 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.586815 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.599563 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.599607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.599619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.599634 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.599700 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.625699 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.666654 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.701648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.701697 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.701707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.701719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.701728 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.804873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.805114 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.805240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.805317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.805393 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.908376 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.908425 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.908438 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.908459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:07 crc kubenswrapper[4758]: I1203 16:56:07.908471 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:07Z","lastTransitionTime":"2025-12-03T16:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.010973 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.011240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.011369 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.011482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.011567 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.113576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.113837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.113900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.113992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.114076 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.216880 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.216913 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.216923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.216939 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.216949 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.280541 4758 generic.go:334] "Generic (PLEG): container finished" podID="06439031-4230-488c-9422-97448eeac39d" containerID="5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd" exitCode=0 Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.280598 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerDied","Data":"5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.311451 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.319306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.319357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.319368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.319386 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.319398 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.324066 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.337935 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.350336 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.364574 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.377609 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.392703 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.411785 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.423884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.423922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.423931 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.423956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.423971 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.427332 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.444294 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.468411 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.483212 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.499777 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.508620 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-c4hd9"] Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.509118 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.511374 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.511667 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.511803 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.514416 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.516708 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.529414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.529440 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.529453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.529466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.529475 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.533919 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.551939 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.567043 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.581296 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.593333 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.605155 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.616025 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.632286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.632330 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.632343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.632361 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.632374 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.636289 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.667993 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.679746 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29bhq\" (UniqueName: \"kubernetes.io/projected/824d1181-466f-47b4-83d7-cc7b79868ff5-kube-api-access-29bhq\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.679806 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/824d1181-466f-47b4-83d7-cc7b79868ff5-host\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.679875 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/824d1181-466f-47b4-83d7-cc7b79868ff5-serviceca\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.711008 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.735133 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.735174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.735184 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.735196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.735206 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.753519 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.781314 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29bhq\" (UniqueName: \"kubernetes.io/projected/824d1181-466f-47b4-83d7-cc7b79868ff5-kube-api-access-29bhq\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.781360 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/824d1181-466f-47b4-83d7-cc7b79868ff5-host\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.781398 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/824d1181-466f-47b4-83d7-cc7b79868ff5-serviceca\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.781416 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/824d1181-466f-47b4-83d7-cc7b79868ff5-host\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.782932 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/824d1181-466f-47b4-83d7-cc7b79868ff5-serviceca\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.789984 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.814453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29bhq\" (UniqueName: \"kubernetes.io/projected/824d1181-466f-47b4-83d7-cc7b79868ff5-kube-api-access-29bhq\") pod \"node-ca-c4hd9\" (UID: \"824d1181-466f-47b4-83d7-cc7b79868ff5\") " pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.829536 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-c4hd9" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.838207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.838241 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.838249 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.838262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.838273 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.851195 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.892175 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.948888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.948952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.948970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.948993 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.949009 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:08Z","lastTransitionTime":"2025-12-03T16:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:08 crc kubenswrapper[4758]: I1203 16:56:08.952072 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.052833 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.052895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.052924 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.053016 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.053098 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.114025 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.114083 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:09 crc kubenswrapper[4758]: E1203 16:56:09.114145 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:09 crc kubenswrapper[4758]: E1203 16:56:09.114197 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.114090 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:09 crc kubenswrapper[4758]: E1203 16:56:09.114269 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.155625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.155707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.155730 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.155752 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.155767 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.259939 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.259989 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.260001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.260027 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.260038 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.290139 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.293517 4758 generic.go:334] "Generic (PLEG): container finished" podID="06439031-4230-488c-9422-97448eeac39d" containerID="e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50" exitCode=0 Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.293582 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerDied","Data":"e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.295249 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-c4hd9" event={"ID":"824d1181-466f-47b4-83d7-cc7b79868ff5","Type":"ContainerStarted","Data":"4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.295284 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-c4hd9" event={"ID":"824d1181-466f-47b4-83d7-cc7b79868ff5","Type":"ContainerStarted","Data":"605372a8e3529158747c13b46a4ba842b6cd58390f36b80c0d4c058079289c05"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.322835 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.337637 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.359465 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.362500 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.362539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.362553 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.362568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.362581 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.375826 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.389846 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.405388 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.419261 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.436758 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.451848 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.467246 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.470182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.470210 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.470221 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.470238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.470250 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.480784 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.498136 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.510668 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.527004 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.542231 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.569108 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.572918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.573139 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.573233 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.573317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.573374 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.608818 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.646876 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.676335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.676623 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.676718 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.676796 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.676906 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.686909 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.726972 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.768283 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.780319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.780414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.780434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.780461 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.780479 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.806284 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.859918 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.883861 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.883929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.883941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.883959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.883972 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.888112 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.928585 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.968668 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.986943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.986997 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.987013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.987033 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:09 crc kubenswrapper[4758]: I1203 16:56:09.987048 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:09Z","lastTransitionTime":"2025-12-03T16:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.007465 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.050739 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.089253 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.089864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.089899 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.089910 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.089928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.089944 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.131213 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.191668 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.191721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.191731 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.191746 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.191755 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.294650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.294789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.294824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.294853 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.294896 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.301487 4758 generic.go:334] "Generic (PLEG): container finished" podID="06439031-4230-488c-9422-97448eeac39d" containerID="1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68" exitCode=0 Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.302432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerDied","Data":"1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.319790 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.336832 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.358367 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.371462 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.386359 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.398146 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.399351 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.399382 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.399392 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.399406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.399417 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.411887 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.446084 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.488513 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.503020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.503061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.503070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.503091 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.503107 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.532800 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.571924 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.605457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.605497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.605510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.605531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.605541 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.610225 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.646491 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.694007 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.708650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.708737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.708751 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.708774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.708800 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.730564 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.802439 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.802565 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.802634 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.802829 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:56:18.802806645 +0000 UTC m=+34.004183506 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.802920 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.802967 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:18.802957979 +0000 UTC m=+34.004334840 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.802882 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.802993 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:18.80298748 +0000 UTC m=+34.004364331 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.811911 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.812099 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.812157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.812215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.812267 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.904168 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.904215 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904331 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904358 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904373 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904424 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:18.904408531 +0000 UTC m=+34.105785392 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904825 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904849 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904859 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:10 crc kubenswrapper[4758]: E1203 16:56:10.904887 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:18.904878013 +0000 UTC m=+34.106254874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.916184 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.916205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.916213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.916227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:10 crc kubenswrapper[4758]: I1203 16:56:10.916236 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:10Z","lastTransitionTime":"2025-12-03T16:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.019865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.019925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.019939 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.019956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.019967 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.114160 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:11 crc kubenswrapper[4758]: E1203 16:56:11.114431 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.114999 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.115046 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:11 crc kubenswrapper[4758]: E1203 16:56:11.115171 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:11 crc kubenswrapper[4758]: E1203 16:56:11.115271 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.122989 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.123073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.123083 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.123182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.123255 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.226046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.226089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.226100 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.226120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.226130 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.311248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.311724 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.314753 4758 generic.go:334] "Generic (PLEG): container finished" podID="06439031-4230-488c-9422-97448eeac39d" containerID="8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239" exitCode=0 Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.314816 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerDied","Data":"8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.329882 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.329927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.329941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.329961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.329976 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.332212 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.348918 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.362243 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.376282 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.381562 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.394107 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.412197 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.424799 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.434817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.434906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.434927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.434985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.435009 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.450079 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.462776 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.481588 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.498180 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.536250 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.538283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.538321 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.538335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.538357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.538372 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.556132 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.579731 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.602416 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.616477 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.627512 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.638608 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.640961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.641012 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.641025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.641054 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.641065 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.656746 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.677501 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.698078 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.710799 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.723760 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.744084 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.744196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.744225 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.744271 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.744307 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.744271 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.760356 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.780282 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.807624 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.847771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.847841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.847856 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.847887 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.847909 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.848025 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.887928 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.930149 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.950040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.950100 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.950112 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.950131 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:11 crc kubenswrapper[4758]: I1203 16:56:11.950145 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:11Z","lastTransitionTime":"2025-12-03T16:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.053398 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.053446 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.053461 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.053484 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.053504 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.157316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.157366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.157379 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.157396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.157606 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.260110 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.260166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.260176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.260196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.260208 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.322277 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.323624 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" event={"ID":"06439031-4230-488c-9422-97448eeac39d","Type":"ContainerStarted","Data":"fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.323703 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.351221 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.352844 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.364074 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.364130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.364140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.364157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.364168 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.367531 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.383448 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.396407 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.416343 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.430545 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.447224 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.464793 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.466603 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.466645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.466657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.466693 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.466717 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.481468 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.493762 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.509138 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.526037 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.539865 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.552348 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.568651 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.570165 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.570243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.570259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.570285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.570303 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.596493 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.614242 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.649932 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.673976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.674052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.674068 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.674094 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.674110 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.688154 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.725230 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.766904 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.777050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.777104 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.777115 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.777134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.777146 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.808692 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.860490 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.880474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.880535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.880555 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.880583 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.880604 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.898952 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.932908 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.965493 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.983662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.983789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.983814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.983849 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:12 crc kubenswrapper[4758]: I1203 16:56:12.983874 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:12Z","lastTransitionTime":"2025-12-03T16:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.004668 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.046719 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.089403 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.089547 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.089610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.089630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.089657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.089722 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.113923 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.114435 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.115119 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.115278 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.115603 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.115774 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.127951 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.197215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.197243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.197251 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.197263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.197271 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.300269 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.300322 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.300332 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.300346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.300357 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.328857 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/0.log" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.333413 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e" exitCode=1 Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.333480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.334669 4758 scope.go:117] "RemoveContainer" containerID="86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.361917 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.390405 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.405360 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.405422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.405438 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.405464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.405483 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.416273 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.435205 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.454405 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.479527 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.495874 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.508824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.508910 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.508933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.508964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.508986 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.530974 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.549332 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.572722 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.590288 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.606876 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.612319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.612357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.612367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.612380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.612392 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.648377 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.692389 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.715237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.715284 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.715297 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.715317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.715330 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.737127 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"informers/externalversions/factory.go:140\\\\nI1203 16:56:12.881741 6034 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:12.882126 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:12.882144 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:12.882164 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:12.882169 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:12.882202 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:12.882213 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:56:12.882219 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:56:12.882224 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:12.882412 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:12.882464 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:12.882487 6034 factory.go:656] Stopping watch factory\\\\nI1203 16:56:12.882508 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:12.882543 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.818409 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.818452 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.818462 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.818481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.818511 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.883985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.884031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.884044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.884059 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.884068 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.904884 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.909298 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.909350 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.909362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.909380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.909393 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.928087 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.932812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.932864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.932878 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.932898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.932910 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.946737 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.951854 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.951901 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.951913 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.951934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.951974 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.968474 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.973419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.973715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.973857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.973910 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.973925 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.993394 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:13 crc kubenswrapper[4758]: E1203 16:56:13.993533 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.995183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.995213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.995222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.995238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:13 crc kubenswrapper[4758]: I1203 16:56:13.995250 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:13Z","lastTransitionTime":"2025-12-03T16:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.097470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.097520 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.097533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.097550 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.097561 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.200017 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.200310 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.200422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.200547 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.200648 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.302716 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.302772 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.302802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.302824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.302835 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.340536 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/0.log" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.347779 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.348024 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.362172 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.389110 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"informers/externalversions/factory.go:140\\\\nI1203 16:56:12.881741 6034 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:12.882126 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:12.882144 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:12.882164 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:12.882169 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:12.882202 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:12.882213 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:56:12.882219 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:56:12.882224 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:12.882412 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:12.882464 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:12.882487 6034 factory.go:656] Stopping watch factory\\\\nI1203 16:56:12.882508 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:12.882543 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.405065 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.405114 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.405129 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.405162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.405177 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.408609 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.423297 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.438958 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.455206 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.467201 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.484105 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.499168 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.508102 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.508374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.508557 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.508751 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.508921 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.509746 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.521374 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.534626 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.558118 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.572645 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.590390 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.613346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.614477 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.614676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.614884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.615077 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.718145 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.718182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.718216 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.718230 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.718240 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.821120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.821216 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.821234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.821261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.821277 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.924211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.924261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.924276 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.924297 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:14 crc kubenswrapper[4758]: I1203 16:56:14.924313 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:14Z","lastTransitionTime":"2025-12-03T16:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.027087 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.027123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.027132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.027147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.027156 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.113747 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.113816 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.113747 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:15 crc kubenswrapper[4758]: E1203 16:56:15.113927 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:15 crc kubenswrapper[4758]: E1203 16:56:15.114066 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:15 crc kubenswrapper[4758]: E1203 16:56:15.114158 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.127125 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.131132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.131200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.131215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.131234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.131248 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.142002 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.167602 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.182952 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.199955 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.212966 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.233766 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"informers/externalversions/factory.go:140\\\\nI1203 16:56:12.881741 6034 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:12.882126 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:12.882144 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:12.882164 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:12.882169 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:12.882202 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:12.882213 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:56:12.882219 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:56:12.882224 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:12.882412 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:12.882464 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:12.882487 6034 factory.go:656] Stopping watch factory\\\\nI1203 16:56:12.882508 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:12.882543 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.233934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.234816 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.235008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.235208 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.235396 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.246257 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.259527 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.271938 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.283237 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.294712 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.307706 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.315962 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.322855 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.333952 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.338660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.338742 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.338757 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.338793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.338808 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.354243 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/1.log" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.355383 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/0.log" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.359725 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5" exitCode=1 Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.359815 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.360267 4758 scope.go:117] "RemoveContainer" containerID="86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.360508 4758 scope.go:117] "RemoveContainer" containerID="f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5" Dec 03 16:56:15 crc kubenswrapper[4758]: E1203 16:56:15.360725 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.361437 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.376287 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.392245 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.409517 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.425669 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.440561 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.442098 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.442125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.442134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.442148 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.442157 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.454633 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.482405 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"informers/externalversions/factory.go:140\\\\nI1203 16:56:12.881741 6034 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:12.882126 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:12.882144 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:12.882164 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:12.882169 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:12.882202 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:12.882213 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:56:12.882219 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:56:12.882224 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:12.882412 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:12.882464 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:12.882487 6034 factory.go:656] Stopping watch factory\\\\nI1203 16:56:12.882508 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:12.882543 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.498986 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.516667 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.533658 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.545266 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.545333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.545344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.545382 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.545395 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.550848 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.562467 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.577408 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.593314 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.606164 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.623641 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.648579 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.648633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.648650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.648671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.648709 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.674140 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.709105 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.741015 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.751483 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.751531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.751543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.751561 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.751575 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.768738 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.811672 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86c8134a847e0e861ee6d1ee404c8e830319f885cceaf8b893563394293fe45e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:13Z\\\",\\\"message\\\":\\\"informers/externalversions/factory.go:140\\\\nI1203 16:56:12.881741 6034 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:12.882126 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:12.882144 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:12.882164 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:12.882169 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:12.882202 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:12.882213 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:56:12.882219 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:56:12.882224 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:12.882412 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:12.882464 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:12.882487 6034 factory.go:656] Stopping watch factory\\\\nI1203 16:56:12.882508 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:12.882543 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.847836 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.854610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.854654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.854666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.854704 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.854721 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.885332 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.928248 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.958256 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.958308 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.958319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.958335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.958344 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:15Z","lastTransitionTime":"2025-12-03T16:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:15 crc kubenswrapper[4758]: I1203 16:56:15.969377 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.012748 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.047436 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.061146 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.061220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.061232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.061248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.061263 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.086711 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.129611 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.163460 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.163498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.163525 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.163541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.163550 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.266739 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.266812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.266824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.266847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.266867 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.365831 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/1.log" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.369224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.369280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.369290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.369303 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.369314 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.370944 4758 scope.go:117] "RemoveContainer" containerID="f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5" Dec 03 16:56:16 crc kubenswrapper[4758]: E1203 16:56:16.371116 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.384597 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.401816 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.423161 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.440741 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.459372 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.472171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.472232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.472250 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.472277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.472297 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.477335 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.494499 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.511382 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.535059 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.557644 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.573943 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.575857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.575930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.575956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.575996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.576022 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.606893 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.651020 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.679394 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.679457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.679476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.679505 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.679526 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.697538 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.733036 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.782478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.782531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.782548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.782582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.782610 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.821110 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l"] Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.822069 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.825140 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.825551 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.854053 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.872258 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.883944 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/afe854a4-7a25-4502-8c41-02b637d14c1c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.884062 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/afe854a4-7a25-4502-8c41-02b637d14c1c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.884090 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/afe854a4-7a25-4502-8c41-02b637d14c1c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.884119 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pvqn\" (UniqueName: \"kubernetes.io/projected/afe854a4-7a25-4502-8c41-02b637d14c1c-kube-api-access-5pvqn\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.885385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.885410 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.885419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.885437 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.885451 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.898151 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.928700 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.969998 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:16Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.984834 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/afe854a4-7a25-4502-8c41-02b637d14c1c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.984906 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/afe854a4-7a25-4502-8c41-02b637d14c1c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.984956 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pvqn\" (UniqueName: \"kubernetes.io/projected/afe854a4-7a25-4502-8c41-02b637d14c1c-kube-api-access-5pvqn\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.984993 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/afe854a4-7a25-4502-8c41-02b637d14c1c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.985518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/afe854a4-7a25-4502-8c41-02b637d14c1c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.985739 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/afe854a4-7a25-4502-8c41-02b637d14c1c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.987100 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.987179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.987200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.987249 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.987270 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:16Z","lastTransitionTime":"2025-12-03T16:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:16 crc kubenswrapper[4758]: I1203 16:56:16.992527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/afe854a4-7a25-4502-8c41-02b637d14c1c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.008804 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.030799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pvqn\" (UniqueName: \"kubernetes.io/projected/afe854a4-7a25-4502-8c41-02b637d14c1c-kube-api-access-5pvqn\") pod \"ovnkube-control-plane-749d76644c-gkj9l\" (UID: \"afe854a4-7a25-4502-8c41-02b637d14c1c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.065276 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.089504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.089539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.089548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.089560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.089569 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.109161 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.114325 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.114333 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.114348 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:17 crc kubenswrapper[4758]: E1203 16:56:17.114456 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:17 crc kubenswrapper[4758]: E1203 16:56:17.114649 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:17 crc kubenswrapper[4758]: E1203 16:56:17.114772 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.142334 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.144178 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.187156 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.192093 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.192131 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.192139 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.192153 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.192162 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.226051 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.268103 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.294302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.294339 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.294347 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.294363 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.294375 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.307206 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.344823 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.374536 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" event={"ID":"afe854a4-7a25-4502-8c41-02b637d14c1c","Type":"ContainerStarted","Data":"31a2305e2ebf398b9cb0be4258c3944be1610393d64ae2dce272a3b962d97b60"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.390003 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.397932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.397974 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.397984 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.398003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.398013 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.426056 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:17Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.501498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.501558 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.501575 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.501600 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.501618 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.604885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.604926 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.604936 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.604950 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.604959 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.707022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.707060 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.707070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.707083 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.707093 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.809633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.809688 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.809700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.809715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.809725 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.911884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.911932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.911948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.911966 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:17 crc kubenswrapper[4758]: I1203 16:56:17.911980 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:17Z","lastTransitionTime":"2025-12-03T16:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.014830 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.014893 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.014906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.014923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.014938 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.117236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.117263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.117272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.117284 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.117293 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.221276 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.221333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.221343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.221362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.221373 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.323516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.323548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.323558 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.323571 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.323581 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.328796 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-55dnk"] Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.329232 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.329284 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.342360 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.355632 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.375893 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.379539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" event={"ID":"afe854a4-7a25-4502-8c41-02b637d14c1c","Type":"ContainerStarted","Data":"154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.379632 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" event={"ID":"afe854a4-7a25-4502-8c41-02b637d14c1c","Type":"ContainerStarted","Data":"6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.394730 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgrhn\" (UniqueName: \"kubernetes.io/projected/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-kube-api-access-tgrhn\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.394980 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.396862 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.417899 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.426031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.426408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.426527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.426632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.426739 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.429793 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.444206 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.457809 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.470299 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.482227 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.496747 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgrhn\" (UniqueName: \"kubernetes.io/projected/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-kube-api-access-tgrhn\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.496852 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.497041 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.497137 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:56:18.997110664 +0000 UTC m=+34.198487555 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.499620 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.509457 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.515442 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgrhn\" (UniqueName: \"kubernetes.io/projected/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-kube-api-access-tgrhn\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.523324 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.530109 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.530138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.530147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.530162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.530174 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.534595 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.543278 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.552405 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.560727 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.570221 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.577975 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.594131 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.606284 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.626675 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.632280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.632342 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.632353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.632368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.632378 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.640561 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.658477 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.673354 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.684857 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.698696 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.726180 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.736026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.736125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.736149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.736184 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.736220 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.752199 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.778422 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.807458 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.827277 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.841022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.841057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.841084 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.841101 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.841113 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.841475 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.852784 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:18Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.899630 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.900054 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:56:34.899967524 +0000 UTC m=+50.101344445 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.900272 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.900380 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.900660 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.900893 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:34.900867058 +0000 UTC m=+50.102243959 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.900663 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:18 crc kubenswrapper[4758]: E1203 16:56:18.901164 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:34.901147235 +0000 UTC m=+50.102524126 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.943642 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.943723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.943736 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.943753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:18 crc kubenswrapper[4758]: I1203 16:56:18.943766 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:18Z","lastTransitionTime":"2025-12-03T16:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.001006 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001195 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001330 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001347 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.001297 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001397 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:35.001381996 +0000 UTC m=+50.202758867 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.001532 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001663 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001737 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:56:20.001721864 +0000 UTC m=+35.203098725 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001848 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001872 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001884 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.001920 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:56:35.001911709 +0000 UTC m=+50.203288570 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.046286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.046342 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.046355 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.046377 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.046389 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.114251 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.114280 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.114442 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.114474 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.114536 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:19 crc kubenswrapper[4758]: E1203 16:56:19.114598 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.149144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.149181 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.149190 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.149203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.149223 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.252615 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.252654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.252666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.252697 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.252708 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.354797 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.354838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.354849 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.354864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.354875 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.458466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.458515 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.458524 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.458548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.458559 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.561568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.561623 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.561639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.561660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.561702 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.665826 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.665902 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.665923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.665952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.665975 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.768837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.768962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.768983 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.769014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.769035 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.871775 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.871830 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.871842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.871859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.871871 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.975186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.975284 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.975305 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.975367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:19 crc kubenswrapper[4758]: I1203 16:56:19.975391 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:19Z","lastTransitionTime":"2025-12-03T16:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.011769 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:20 crc kubenswrapper[4758]: E1203 16:56:20.012185 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:20 crc kubenswrapper[4758]: E1203 16:56:20.012329 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:56:22.012292847 +0000 UTC m=+37.213669878 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.078222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.078275 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.078286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.078305 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.078319 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.114289 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:20 crc kubenswrapper[4758]: E1203 16:56:20.114505 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.181405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.181841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.181867 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.181903 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.181928 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.284945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.284995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.285006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.285025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.285036 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.387538 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.387619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.387633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.387655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.387666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.489935 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.490001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.490015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.490029 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.490039 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.593197 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.593263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.593294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.593334 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.593356 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.697121 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.697177 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.697189 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.697207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.697219 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.800283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.800391 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.800411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.800456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.800474 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.903459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.903496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.903506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.903519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:20 crc kubenswrapper[4758]: I1203 16:56:20.903530 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:20Z","lastTransitionTime":"2025-12-03T16:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.006171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.006206 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.006215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.006231 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.006241 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.108567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.108657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.108707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.108940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.108960 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.113881 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.113860 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:21 crc kubenswrapper[4758]: E1203 16:56:21.114091 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.113906 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:21 crc kubenswrapper[4758]: E1203 16:56:21.114157 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:21 crc kubenswrapper[4758]: E1203 16:56:21.114261 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.212295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.212345 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.212353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.212368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.212378 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.315619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.315661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.315669 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.315706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.315716 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.417975 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.418014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.418024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.418037 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.418046 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.520425 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.520459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.520468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.520480 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.520489 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.622740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.623013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.623077 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.623134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.623195 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.726145 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.726398 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.726460 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.726561 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.726638 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.829607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.829741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.829773 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.829810 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.829839 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.932610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.932672 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.932700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.932717 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:21 crc kubenswrapper[4758]: I1203 16:56:21.932727 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:21Z","lastTransitionTime":"2025-12-03T16:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.031551 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:22 crc kubenswrapper[4758]: E1203 16:56:22.031753 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:22 crc kubenswrapper[4758]: E1203 16:56:22.031862 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:56:26.031840271 +0000 UTC m=+41.233217202 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.035204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.035258 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.035270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.035289 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.035303 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.113765 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:22 crc kubenswrapper[4758]: E1203 16:56:22.113956 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.139218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.139300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.139329 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.139368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.139395 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.242765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.243365 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.243583 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.243810 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.244013 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.347894 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.348127 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.348146 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.348176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.348195 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.451992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.452527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.452670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.452857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.452994 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.556153 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.556224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.556241 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.556267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.556286 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.660371 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.660897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.661331 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.662084 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.662305 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.765316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.765362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.765372 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.765388 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.765399 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.873117 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.873179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.873193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.873214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.873228 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.976085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.976140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.976152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.976169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:22 crc kubenswrapper[4758]: I1203 16:56:22.976184 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:22Z","lastTransitionTime":"2025-12-03T16:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.078907 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.078962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.078973 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.078989 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.079000 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.113852 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:23 crc kubenswrapper[4758]: E1203 16:56:23.113974 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.113993 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:23 crc kubenswrapper[4758]: E1203 16:56:23.114106 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.114145 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:23 crc kubenswrapper[4758]: E1203 16:56:23.114386 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.181095 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.181142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.181156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.181175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.181190 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.285542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.285616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.285637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.285823 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.285869 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.388494 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.388552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.388569 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.388601 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.388620 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.491661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.491723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.491736 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.491754 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.491767 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.595106 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.595195 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.595220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.595288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.595312 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.698885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.698942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.698959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.698984 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.699000 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.801648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.801679 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.801687 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.801700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.801708 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.905053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.905149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.905169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.905192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:23 crc kubenswrapper[4758]: I1203 16:56:23.905246 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:23Z","lastTransitionTime":"2025-12-03T16:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.007955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.008006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.008018 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.008036 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.008047 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.111424 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.111469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.111481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.111498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.111508 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.113825 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.113956 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.196602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.196644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.196655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.196673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.196728 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.215579 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:24Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.219509 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.219539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.219552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.219570 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.219580 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.232820 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:24Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.237904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.237930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.237940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.237952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.237961 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.249855 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:24Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.254155 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.254186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.254196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.254211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.254221 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.266595 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:24Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.270489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.270539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.270551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.270572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.270584 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.283229 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:24Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:24 crc kubenswrapper[4758]: E1203 16:56:24.283394 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.285406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.285464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.285477 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.285498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.285513 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.388505 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.388555 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.388563 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.388578 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.388588 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.492464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.492686 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.492751 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.492791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.492815 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.595632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.595680 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.595692 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.595727 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.595740 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.698154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.698190 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.698200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.698213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.698222 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.800103 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.800142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.800152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.800167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.800178 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.904250 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.904340 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.904362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.904389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:24 crc kubenswrapper[4758]: I1203 16:56:24.904410 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:24Z","lastTransitionTime":"2025-12-03T16:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.006365 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.006408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.006417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.006431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.006440 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.108289 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.108337 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.108348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.108363 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.108374 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.113565 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.113577 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:25 crc kubenswrapper[4758]: E1203 16:56:25.113768 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.113794 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:25 crc kubenswrapper[4758]: E1203 16:56:25.113878 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:25 crc kubenswrapper[4758]: E1203 16:56:25.114045 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.131497 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.143584 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.157329 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.174765 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.196895 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.210133 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.210817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.210851 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.210862 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.210879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.210891 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.225310 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.241387 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.253880 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.267335 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.277530 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.290103 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.300863 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.318449 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.318490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.318503 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.318519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.318530 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.318536 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.329489 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.343472 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.352580 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:25Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.421126 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.421179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.421187 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.421199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.421208 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.523348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.523396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.523410 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.523428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.523835 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.626447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.626485 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.626498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.626527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.626539 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.729440 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.729488 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.729499 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.729516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.729530 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.832429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.832485 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.832519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.832541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.832551 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.935374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.935419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.935430 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.935451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:25 crc kubenswrapper[4758]: I1203 16:56:25.935498 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:25Z","lastTransitionTime":"2025-12-03T16:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.038000 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.038074 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.038109 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.038140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.038165 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.079893 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:26 crc kubenswrapper[4758]: E1203 16:56:26.080110 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:26 crc kubenswrapper[4758]: E1203 16:56:26.080191 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:56:34.080171788 +0000 UTC m=+49.281548649 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.114148 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:26 crc kubenswrapper[4758]: E1203 16:56:26.114332 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.140615 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.140713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.140725 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.140742 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.140754 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.243048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.243266 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.243345 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.243452 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.243523 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.346802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.346841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.346851 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.346864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.346876 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.450295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.450354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.450365 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.450382 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.450393 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.553393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.553456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.553466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.553480 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.553490 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.655925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.655965 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.655974 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.655988 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.655999 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.757841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.757874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.757885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.757904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.757916 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.859892 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.859921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.859931 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.859945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.859954 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.962598 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.962703 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.962718 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.962734 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:26 crc kubenswrapper[4758]: I1203 16:56:26.962745 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:26Z","lastTransitionTime":"2025-12-03T16:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.065586 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.065643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.065661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.065721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.065739 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.114092 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.114167 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:27 crc kubenswrapper[4758]: E1203 16:56:27.114250 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.114126 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:27 crc kubenswrapper[4758]: E1203 16:56:27.114355 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:27 crc kubenswrapper[4758]: E1203 16:56:27.114423 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.171514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.171559 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.171784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.171805 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.171816 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.274625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.275003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.275121 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.275203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.275284 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.377472 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.377517 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.377526 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.377540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.377549 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.480407 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.480458 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.480471 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.480489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.480501 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.583104 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.583147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.583158 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.583174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.583188 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.686329 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.686367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.686376 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.686392 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.686401 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.788624 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.788755 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.788777 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.788796 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.788809 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.891007 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.891056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.891073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.891095 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.891113 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.993316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.993353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.993362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.993374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:27 crc kubenswrapper[4758]: I1203 16:56:27.993382 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:27Z","lastTransitionTime":"2025-12-03T16:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.096443 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.096481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.096492 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.096507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.096516 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.114121 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:28 crc kubenswrapper[4758]: E1203 16:56:28.114272 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.115122 4758 scope.go:117] "RemoveContainer" containerID="f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.199542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.199580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.199596 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.199618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.199632 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.301813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.301876 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.301887 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.301903 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.301915 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.403499 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.403565 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.403580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.403598 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.403627 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.412343 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/1.log" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.415145 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.415388 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.430658 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.458664 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.477392 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.496819 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.506259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.506310 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.506320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.506339 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.506352 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.513316 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.533544 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.548297 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.567291 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.585787 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.606080 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.615157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.615200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.615209 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.615228 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.615240 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.626755 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.646886 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.669131 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.696733 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.710594 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.717525 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.717551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.717558 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.717572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.717582 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.723098 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.736335 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:28Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.820115 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.820153 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.820163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.820179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.820189 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.922987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.923023 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.923031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.923044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:28 crc kubenswrapper[4758]: I1203 16:56:28.923053 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:28Z","lastTransitionTime":"2025-12-03T16:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.025521 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.025572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.025586 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.025603 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.025613 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.113265 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.113401 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:29 crc kubenswrapper[4758]: E1203 16:56:29.113482 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.113556 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:29 crc kubenswrapper[4758]: E1203 16:56:29.113649 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:29 crc kubenswrapper[4758]: E1203 16:56:29.113815 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.127788 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.127816 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.127825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.127838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.127848 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.230398 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.230437 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.230449 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.230466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.230477 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.332573 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.332614 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.332625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.332641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.332653 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.435602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.435654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.435666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.435711 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.435725 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.537883 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.537919 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.537927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.537942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.537952 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.641283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.641322 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.641333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.641345 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.641355 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.744102 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.744148 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.744161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.744178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.744189 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.846560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.846596 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.846606 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.846620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.846630 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.950303 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.950349 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.950361 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.950381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:29 crc kubenswrapper[4758]: I1203 16:56:29.950397 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:29Z","lastTransitionTime":"2025-12-03T16:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.053287 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.053353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.053366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.053384 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.053396 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.113368 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:30 crc kubenswrapper[4758]: E1203 16:56:30.113543 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.155927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.155986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.155999 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.156014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.156024 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.259114 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.259154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.259165 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.259180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.259210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.362276 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.362318 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.362328 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.362353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.362362 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.422478 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/2.log" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.423019 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/1.log" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.425876 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955" exitCode=1 Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.425915 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.426019 4758 scope.go:117] "RemoveContainer" containerID="f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.426595 4758 scope.go:117] "RemoveContainer" containerID="5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955" Dec 03 16:56:30 crc kubenswrapper[4758]: E1203 16:56:30.426790 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.441993 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.456363 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.464457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.464478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.464485 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.464498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.464507 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.468077 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.477413 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.486771 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.496740 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.504807 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.524996 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.538099 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.553485 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.564159 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.566620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.566647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.566658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.566675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.566707 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.577760 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.590139 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.603038 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.615790 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.634525 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64594c6f531b60cf2f92187c073cdd946832c9f18ebed4640693658212a93e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:14Z\\\",\\\"message\\\":\\\"flector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:14.258456 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 16:56:14.258476 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:56:14.258494 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:56:14.258507 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:56:14.258525 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:56:14.258539 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:56:14.258561 6174 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:56:14.258565 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 16:56:14.258583 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:56:14.258627 6174 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 16:56:14.258644 6174 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 16:56:14.258715 6174 factory.go:656] Stopping watch factory\\\\nI1203 16:56:14.258723 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:56:14.258741 6174 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 16:56:14.258744 6174 ovnkube.go:599] Stopped ovnkube\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.645989 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.668873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.668907 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.668916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.668928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.668937 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.771360 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.771417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.771427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.771444 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.771790 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.873881 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.874121 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.874193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.874288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.874344 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.976875 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.977168 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.977239 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.977324 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:30 crc kubenswrapper[4758]: I1203 16:56:30.977407 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:30Z","lastTransitionTime":"2025-12-03T16:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.079848 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.080223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.080295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.080514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.080665 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.113775 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.114067 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:31 crc kubenswrapper[4758]: E1203 16:56:31.114201 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:31 crc kubenswrapper[4758]: E1203 16:56:31.114257 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.114425 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:31 crc kubenswrapper[4758]: E1203 16:56:31.114597 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.182950 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.183509 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.183653 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.183814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.183897 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.286916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.286960 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.286970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.286985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.286995 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.390373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.390419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.390430 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.390448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.390460 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.431831 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/2.log" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.494312 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.494351 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.494364 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.494385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.494397 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.596681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.596725 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.596733 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.596745 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.596754 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.698582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.698617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.698626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.698639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.698648 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.802265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.802567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.802649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.802779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.802860 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.912091 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.912133 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.912146 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.912174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:31 crc kubenswrapper[4758]: I1203 16:56:31.912204 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:31Z","lastTransitionTime":"2025-12-03T16:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.014633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.014672 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.014734 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.014753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.014764 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.113749 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:32 crc kubenswrapper[4758]: E1203 16:56:32.113953 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.117407 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.117440 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.117449 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.117462 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.117471 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.220214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.220284 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.220296 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.220547 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.220559 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.323531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.323566 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.323577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.323594 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.323606 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.425681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.425733 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.425740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.425753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.425764 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.527616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.527926 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.528031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.528120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.528202 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.629882 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.629929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.629940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.629953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.629961 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.732509 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.732561 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.732571 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.732587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.732598 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.834786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.835063 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.835130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.835201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.835267 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.937554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.937587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.937595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.937608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:32 crc kubenswrapper[4758]: I1203 16:56:32.937617 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:32Z","lastTransitionTime":"2025-12-03T16:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.004541 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.005333 4758 scope.go:117] "RemoveContainer" containerID="5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955" Dec 03 16:56:33 crc kubenswrapper[4758]: E1203 16:56:33.005498 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.021436 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.038927 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.039589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.039670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.039828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.039863 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.039887 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.052971 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.066921 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.078431 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.091486 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.102332 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.114666 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:33 crc kubenswrapper[4758]: E1203 16:56:33.114859 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.115096 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:33 crc kubenswrapper[4758]: E1203 16:56:33.115471 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.115411 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:33 crc kubenswrapper[4758]: E1203 16:56:33.115722 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.126089 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.140622 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.141564 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.141666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.141747 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.141817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.141978 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.158234 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.171425 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.185491 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.199764 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.214026 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.230272 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.245175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.245231 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.245249 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.245268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.245280 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.250270 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.261124 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.352975 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.353051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.353075 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.353102 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.353125 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.455524 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.455587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.455605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.455629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.455645 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.558414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.558482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.558508 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.558534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.558552 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.661366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.661428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.661448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.661476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.661494 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.784779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.784822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.784834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.784850 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.784862 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.887004 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.887043 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.887052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.887066 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.887075 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.990651 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.990768 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.990794 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.990829 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:33 crc kubenswrapper[4758]: I1203 16:56:33.990852 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:33Z","lastTransitionTime":"2025-12-03T16:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.088437 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.088592 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.088636 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:56:50.088623606 +0000 UTC m=+65.290000467 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.092787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.092817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.092827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.092843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.092854 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.114368 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.114927 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.195671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.195805 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.195836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.195876 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.195902 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.299893 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.299980 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.300020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.300062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.300076 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.403147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.403230 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.403248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.403279 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.403296 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.505835 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.506127 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.506208 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.506283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.506356 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.578392 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.578433 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.578442 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.578459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.578469 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.590091 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:34Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.594291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.594335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.594367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.594384 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.594394 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.609133 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:34Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.614067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.614110 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.614122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.614147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.614161 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.631876 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:34Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.636777 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.636815 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.636827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.636846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.636858 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.647806 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:34Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.652644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.652846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.652933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.653018 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.653085 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.664945 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:34Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.665063 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.666939 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.666991 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.667001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.667020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.667031 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.770157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.770249 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.770270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.770300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.770321 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.873423 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.873955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.874113 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.874252 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.874400 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.977397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.977427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.977436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.977450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.977460 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:34Z","lastTransitionTime":"2025-12-03T16:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.998595 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.998832 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:34 crc kubenswrapper[4758]: I1203 16:56:34.998871 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.998984 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.999052 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:06.99903261 +0000 UTC m=+82.200409481 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.999359 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.999377 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:06.99936089 +0000 UTC m=+82.200737761 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:56:34 crc kubenswrapper[4758]: E1203 16:56:34.999438 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:06.999427411 +0000 UTC m=+82.200804282 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.079767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.080002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.080067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.080127 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.080185 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.100306 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.100612 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.100838 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.100920 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.100988 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.101078 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:07.101065327 +0000 UTC m=+82.302442188 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.100848 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.101197 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.101210 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.101245 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:07.101235532 +0000 UTC m=+82.302612393 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.113806 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.114029 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.114029 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.114165 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.114419 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:35 crc kubenswrapper[4758]: E1203 16:56:35.114558 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.123852 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.135968 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.149099 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.168453 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.180657 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.182342 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.182401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.182419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.182448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.182467 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.195610 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.210858 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.241450 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.254790 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.273070 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.284470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.284523 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.284532 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.284549 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.284561 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.286251 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.302018 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.318767 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.332891 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.350195 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.367177 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.379647 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:35Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.387497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.387532 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.387542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.387564 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.387576 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.490041 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.490142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.490172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.490207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.490234 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.593134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.593203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.593222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.593248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.593272 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.696752 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.696836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.696855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.696889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.696911 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.799773 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.799847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.799858 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.799895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.799907 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.903425 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.903473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.903482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.903497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:35 crc kubenswrapper[4758]: I1203 16:56:35.903508 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:35Z","lastTransitionTime":"2025-12-03T16:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.007404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.007476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.007494 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.007523 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.007542 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.110843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.110902 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.110912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.110933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.110947 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.114050 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:36 crc kubenswrapper[4758]: E1203 16:56:36.114181 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.213782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.213812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.213820 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.213833 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.213842 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.316104 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.316139 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.316149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.316164 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.316174 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.418043 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.418109 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.418120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.418136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.418146 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.520778 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.520824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.520834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.520853 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.520864 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.622947 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.622995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.623010 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.623029 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.623042 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.725947 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.725992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.726004 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.726019 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.726029 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.828979 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.829024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.829038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.829057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.829071 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.931836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.931884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.931895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.931910 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:36 crc kubenswrapper[4758]: I1203 16:56:36.931921 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:36Z","lastTransitionTime":"2025-12-03T16:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.033726 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.033776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.033791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.033814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.033831 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.113932 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.113977 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.114005 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:37 crc kubenswrapper[4758]: E1203 16:56:37.114088 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:37 crc kubenswrapper[4758]: E1203 16:56:37.114169 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:37 crc kubenswrapper[4758]: E1203 16:56:37.114256 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.135469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.135507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.135520 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.135537 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.135550 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.237888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.237923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.237932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.237946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.237955 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.340700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.340733 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.340741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.340755 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.340766 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.443916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.443964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.443976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.443993 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.444003 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.546547 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.546574 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.546582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.546596 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.546607 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.648791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.648837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.648848 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.648867 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.648879 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.751637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.751670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.751697 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.751711 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.751727 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.855219 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.855278 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.855290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.855313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.855326 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.959255 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.959309 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.959322 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.959344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:37 crc kubenswrapper[4758]: I1203 16:56:37.959355 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:37Z","lastTransitionTime":"2025-12-03T16:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.062585 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.062919 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.063005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.063079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.063154 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.113395 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:38 crc kubenswrapper[4758]: E1203 16:56:38.113809 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.166121 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.166171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.166194 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.166222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.166241 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.268889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.268935 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.268946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.268962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.268974 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.372019 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.372090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.372101 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.372116 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.372127 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.473963 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.474033 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.474044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.474062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.474071 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.576786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.577255 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.577466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.577649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.577828 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.681674 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.681993 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.682076 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.682174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.682256 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.784451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.784489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.784500 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.784514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.784524 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.887733 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.887792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.887809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.887831 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.887847 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.990753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.990796 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.990807 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.990823 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:38 crc kubenswrapper[4758]: I1203 16:56:38.990833 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:38Z","lastTransitionTime":"2025-12-03T16:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.093228 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.093317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.093335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.093368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.093388 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.113899 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.113907 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.114142 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:39 crc kubenswrapper[4758]: E1203 16:56:39.114261 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:39 crc kubenswrapper[4758]: E1203 16:56:39.114085 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:39 crc kubenswrapper[4758]: E1203 16:56:39.114507 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.196258 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.196307 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.196319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.196337 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.196350 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.299380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.299461 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.299486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.299516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.299536 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.402875 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.402943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.402961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.402987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.403004 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.506368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.506765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.506916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.507082 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.507221 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.610820 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.610905 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.610927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.610959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.610980 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.714422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.714906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.715103 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.715272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.715421 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.819760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.819831 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.819846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.819873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.819893 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.922534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.922620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.922635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.922656 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:39 crc kubenswrapper[4758]: I1203 16:56:39.922672 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:39Z","lastTransitionTime":"2025-12-03T16:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.026112 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.026171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.026182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.026199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.026210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.113310 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:40 crc kubenswrapper[4758]: E1203 16:56:40.113497 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.129070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.129142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.129156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.129189 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.129209 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.232183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.232845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.232867 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.232886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.232899 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.336884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.336940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.336976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.337003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.337021 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.439861 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.440397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.440459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.440797 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.440920 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.544270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.544558 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.544663 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.544787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.544885 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.647832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.648080 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.648145 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.648221 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.648277 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.684049 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.698836 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.702507 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.738057 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.751613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.751724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.751752 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.751786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.751806 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.759299 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.780893 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.798524 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.832354 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.854159 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.855337 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.855414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.855439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.855474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.855500 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.875586 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.899085 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.919912 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.936853 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.951988 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.958411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.958480 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.958501 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.958531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.958556 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:40Z","lastTransitionTime":"2025-12-03T16:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:40 crc kubenswrapper[4758]: I1203 16:56:40.978524 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:40Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.003719 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.023304 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.041553 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.062294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.062671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.062712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.062738 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.062754 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.063319 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.114007 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.114007 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:41 crc kubenswrapper[4758]: E1203 16:56:41.114202 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:41 crc kubenswrapper[4758]: E1203 16:56:41.114350 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.114592 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:41 crc kubenswrapper[4758]: E1203 16:56:41.114893 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.165269 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.165534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.165595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.165665 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.165751 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.269261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.269343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.269360 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.269389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.269411 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.372523 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.372584 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.372602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.372630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.372650 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.476350 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.476452 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.476473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.476525 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.476545 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.579397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.579452 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.579464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.579482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.579493 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.683009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.683103 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.683116 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.683136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.683153 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.787010 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.787094 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.787114 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.787141 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.787159 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.890226 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.890778 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.891014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.891235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.891636 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.995567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.996173 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.996315 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.996477 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:41 crc kubenswrapper[4758]: I1203 16:56:41.996571 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:41Z","lastTransitionTime":"2025-12-03T16:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.099719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.100051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.100142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.100262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.100352 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.113414 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:42 crc kubenswrapper[4758]: E1203 16:56:42.113599 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.202535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.202588 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.202600 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.202620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.202633 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.306062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.306119 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.306133 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.306157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.306170 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.408922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.409211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.409317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.409399 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.409482 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.512424 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.512520 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.512540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.512574 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.512595 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.615998 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.616070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.616090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.616128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.616154 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.719308 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.719372 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.719383 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.719402 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.719416 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.822575 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.822623 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.822634 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.822648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.822659 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.926052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.926144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.926163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.926196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:42 crc kubenswrapper[4758]: I1203 16:56:42.926220 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:42Z","lastTransitionTime":"2025-12-03T16:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.028706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.028749 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.028761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.028776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.028804 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.114371 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.114410 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.114469 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:43 crc kubenswrapper[4758]: E1203 16:56:43.114585 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:43 crc kubenswrapper[4758]: E1203 16:56:43.114894 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:43 crc kubenswrapper[4758]: E1203 16:56:43.115130 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.130898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.130937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.130948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.130963 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.130972 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.233357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.233401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.233412 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.233431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.233442 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.336018 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.336064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.336077 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.336095 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.336107 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.438428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.438706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.438826 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.438924 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.439014 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.541841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.541892 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.541900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.541912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.541920 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.644660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.644730 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.644741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.644756 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.644767 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.747445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.747746 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.747868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.747976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.748058 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.850988 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.851049 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.851064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.851088 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.851104 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.954468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.954740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.954812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.954880 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:43 crc kubenswrapper[4758]: I1203 16:56:43.954942 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:43Z","lastTransitionTime":"2025-12-03T16:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.058576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.058928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.059021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.059129 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.059221 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.113548 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.114879 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.114778 4758 scope.go:117] "RemoveContainer" containerID="5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955" Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.115297 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.163369 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.163719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.163837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.163989 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.164123 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.266860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.266925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.266937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.266960 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.266973 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.370836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.370898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.370911 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.370932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.370946 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.474865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.474917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.474930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.474949 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.474962 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.580062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.580120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.580135 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.580160 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.580178 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.683877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.683927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.683940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.683968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.683981 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.787957 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.788032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.788048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.788070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.788085 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.825606 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.825720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.825741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.825774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.825793 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.839461 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:44Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.844791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.844976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.845082 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.845159 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.845229 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.860832 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:44Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.870186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.870268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.870280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.870302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.870368 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.885571 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:44Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.891986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.892073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.892097 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.892131 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.892159 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.913449 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:44Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.918618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.918760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.918792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.918832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.918860 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.942791 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:44Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:44 crc kubenswrapper[4758]: E1203 16:56:44.943152 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.944936 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.944972 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.944982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.944997 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:44 crc kubenswrapper[4758]: I1203 16:56:44.945007 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:44Z","lastTransitionTime":"2025-12-03T16:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.047554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.047592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.047602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.047617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.047628 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.113330 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.113459 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.113505 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:45 crc kubenswrapper[4758]: E1203 16:56:45.114105 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:45 crc kubenswrapper[4758]: E1203 16:56:45.114338 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:45 crc kubenswrapper[4758]: E1203 16:56:45.114533 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.134149 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.150265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.150305 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.150315 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.150331 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.150341 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.155144 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.187139 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.206160 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.226138 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.244497 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.254196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.254307 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.255831 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.255904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.255932 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.261839 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.279967 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.294542 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.314399 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.329329 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.343022 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.355923 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.359208 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.359246 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.359259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.359278 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.359290 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.369879 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.391233 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.404532 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.452112 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.461519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.461565 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.461575 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.461593 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.461606 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.481395 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:45Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.565529 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.565584 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.565597 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.565615 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.565626 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.668660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.668721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.668731 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.668750 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.668760 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.771878 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.771919 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.771929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.771943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.771952 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.874326 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.874389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.874404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.874456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.874478 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.977005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.977061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.977072 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.977092 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:45 crc kubenswrapper[4758]: I1203 16:56:45.977106 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:45Z","lastTransitionTime":"2025-12-03T16:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.080637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.081089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.081103 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.081128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.081142 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.113409 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:46 crc kubenswrapper[4758]: E1203 16:56:46.113542 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.184542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.184625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.184650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.184713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.184735 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.287841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.287895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.287904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.287922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.287935 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.396271 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.396344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.396365 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.396390 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.396409 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.500928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.501024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.501053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.501086 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.501111 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.603925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.603996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.604039 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.604079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.604108 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.706567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.706609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.706617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.706632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.706641 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.810037 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.810114 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.810130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.810149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.810165 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.912602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.912641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.912650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.912663 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:46 crc kubenswrapper[4758]: I1203 16:56:46.912674 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:46Z","lastTransitionTime":"2025-12-03T16:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.014849 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.014893 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.014901 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.014917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.014927 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.113381 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.113403 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:47 crc kubenswrapper[4758]: E1203 16:56:47.113543 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.113404 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:47 crc kubenswrapper[4758]: E1203 16:56:47.113647 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:47 crc kubenswrapper[4758]: E1203 16:56:47.113969 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.117069 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.117123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.117140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.117160 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.117171 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.220046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.220096 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.220108 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.220130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.220144 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.322835 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.322931 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.322950 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.322973 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.322989 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.425429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.425464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.425474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.425487 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.425497 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.528537 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.528611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.528621 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.528637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.528646 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.631956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.632042 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.632061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.632090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.632116 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.735320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.735362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.735373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.735390 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.735401 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.839192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.839236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.839246 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.839265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.839276 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.941737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.941780 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.941791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.941807 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:47 crc kubenswrapper[4758]: I1203 16:56:47.941818 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:47Z","lastTransitionTime":"2025-12-03T16:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.045004 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.045067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.045079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.045106 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.045118 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.114184 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:48 crc kubenswrapper[4758]: E1203 16:56:48.114352 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.149969 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.150122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.150142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.150161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.150174 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.252204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.252252 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.252261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.252273 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.252283 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.355088 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.355127 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.355135 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.355150 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.355159 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.456862 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.456904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.456914 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.456929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.456941 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.559427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.559471 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.559494 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.559510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.559520 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.662233 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.662279 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.662293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.662311 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.662323 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.764747 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.764791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.764800 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.764841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.764854 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.867976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.868019 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.868028 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.868045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.868056 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.973540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.973651 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.973699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.973728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:48 crc kubenswrapper[4758]: I1203 16:56:48.973747 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:48Z","lastTransitionTime":"2025-12-03T16:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.076771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.076821 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.076834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.076854 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.076868 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.113586 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.113587 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.113607 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:49 crc kubenswrapper[4758]: E1203 16:56:49.113771 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:49 crc kubenswrapper[4758]: E1203 16:56:49.113877 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:49 crc kubenswrapper[4758]: E1203 16:56:49.113960 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.180241 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.180290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.180303 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.180322 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.180337 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.283496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.283529 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.283538 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.283551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.283562 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.385722 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.385759 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.385769 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.385783 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.385793 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.489099 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.489153 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.489163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.489182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.489194 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.591724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.591781 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.591798 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.591815 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.592010 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.695344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.695376 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.695387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.695404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.695414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.797819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.797870 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.797883 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.797901 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.797912 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.901116 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.901168 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.901183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.901205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:49 crc kubenswrapper[4758]: I1203 16:56:49.901221 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:49Z","lastTransitionTime":"2025-12-03T16:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.004331 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.004717 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.004812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.004962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.005055 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.108172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.108223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.108235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.108287 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.108302 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.113562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:50 crc kubenswrapper[4758]: E1203 16:56:50.113768 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.175044 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:50 crc kubenswrapper[4758]: E1203 16:56:50.175252 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:50 crc kubenswrapper[4758]: E1203 16:56:50.175344 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:57:22.175318448 +0000 UTC m=+97.376695309 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.211325 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.211383 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.211399 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.211425 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.211443 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.314953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.315014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.315026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.315044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.315058 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.418618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.418707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.418724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.418742 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.418753 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.521160 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.521200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.521209 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.521223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.521231 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.625044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.625125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.625138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.625163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.625178 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.728281 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.728743 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.728838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.728930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.729017 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.833069 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.833122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.833140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.833158 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.833171 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.936159 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.936211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.936222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.936239 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:50 crc kubenswrapper[4758]: I1203 16:56:50.936254 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:50Z","lastTransitionTime":"2025-12-03T16:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.038888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.038967 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.038980 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.039006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.039023 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.113964 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.114060 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.113981 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:51 crc kubenswrapper[4758]: E1203 16:56:51.114125 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:51 crc kubenswrapper[4758]: E1203 16:56:51.114261 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:51 crc kubenswrapper[4758]: E1203 16:56:51.114416 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.141320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.141380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.141395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.141411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.141431 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.245671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.245735 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.245747 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.245765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.245775 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.348876 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.348933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.348944 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.348965 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.348977 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.451662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.451732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.451741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.451755 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.451765 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.500907 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/0.log" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.500976 4758 generic.go:334] "Generic (PLEG): container finished" podID="4b5429d1-1fc3-4603-93ba-b57b33c2f585" containerID="3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb" exitCode=1 Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.501017 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerDied","Data":"3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.501489 4758 scope.go:117] "RemoveContainer" containerID="3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.516747 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.530657 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.544030 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.554211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.554265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.554278 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.554296 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.554307 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.579796 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.593382 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.608639 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.619276 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.631532 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.644536 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.657375 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.657413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.657425 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.657441 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.657453 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.659209 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.674771 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.689880 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.706983 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.719794 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.733015 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.745451 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.759996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.760066 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.760083 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.760108 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.760122 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.761808 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.775667 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.862779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.862833 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.862843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.862861 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.862871 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.966080 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.966144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.966158 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.966181 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:51 crc kubenswrapper[4758]: I1203 16:56:51.966196 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:51Z","lastTransitionTime":"2025-12-03T16:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.069816 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.069879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.069891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.069909 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.069920 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.113874 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:52 crc kubenswrapper[4758]: E1203 16:56:52.114023 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.172391 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.172456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.172471 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.172491 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.172503 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.275441 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.275524 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.275535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.275552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.275565 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.377776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.377853 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.377864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.377886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.377909 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.480455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.480499 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.480513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.480533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.480547 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.504975 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/0.log" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.505040 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerStarted","Data":"949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.521615 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.532807 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.544718 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.560792 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.574337 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.583469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.583532 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.583549 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.583576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.583589 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.587257 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.599940 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.613409 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.636823 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.651049 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.675828 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.686993 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.687020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.687031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.687046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.687056 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.688566 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.700553 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.713719 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.734267 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.749667 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.763739 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.777856 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:52Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.789434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.789456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.789463 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.789476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.789484 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.892416 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.892468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.892478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.892497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.892509 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.995044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.995081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.995089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.995104 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:52 crc kubenswrapper[4758]: I1203 16:56:52.995114 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:52Z","lastTransitionTime":"2025-12-03T16:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.096630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.096664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.096675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.096702 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.096712 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.113405 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:53 crc kubenswrapper[4758]: E1203 16:56:53.113525 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.113520 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.113574 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:53 crc kubenswrapper[4758]: E1203 16:56:53.113628 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:53 crc kubenswrapper[4758]: E1203 16:56:53.113842 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.199103 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.199149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.199162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.199179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.199193 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.301751 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.301803 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.301818 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.301837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.301851 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.405581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.405654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.405666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.405714 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.405730 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.507707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.507748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.507761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.507780 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.507792 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.610428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.610475 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.610487 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.610504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.610516 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.713456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.713504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.713516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.713536 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.713549 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.815908 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.816005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.816027 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.816056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.816077 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.919048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.919096 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.919107 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.919126 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:53 crc kubenswrapper[4758]: I1203 16:56:53.919137 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:53Z","lastTransitionTime":"2025-12-03T16:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.021840 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.021888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.021900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.021917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.021930 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.113490 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:54 crc kubenswrapper[4758]: E1203 16:56:54.113645 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.124013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.124067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.124094 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.124117 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.124134 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.227117 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.227157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.227165 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.227183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.227194 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.329668 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.329737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.329752 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.329771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.329786 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.432482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.432837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.432847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.432860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.432869 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.534933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.534980 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.534999 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.535020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.535037 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.673166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.673281 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.673297 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.673312 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.673323 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.776467 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.776562 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.776581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.776605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.776622 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.879626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.879727 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.879747 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.879780 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.879807 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.981567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.981609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.981618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.981632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:54 crc kubenswrapper[4758]: I1203 16:56:54.981641 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:54Z","lastTransitionTime":"2025-12-03T16:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.085668 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.085786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.085802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.085824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.085841 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.113815 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.113977 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.114040 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.114087 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.114220 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.114378 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.132370 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.149034 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.166190 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.182406 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.188845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.188900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.188917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.188944 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.188962 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.190730 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.190779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.190804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.190832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.190854 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.201182 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.212671 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.217493 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.217767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.217954 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.218167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.218260 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.222702 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.232259 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.237556 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.239398 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.239564 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.239657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.239767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.239860 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.257172 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.259604 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.265507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.265560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.265573 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.265594 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.265608 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.273183 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.278768 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.293178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.293447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.293556 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.293657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.293794 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.293508 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.311278 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.312401 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: E1203 16:56:55.312514 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.315413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.315471 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.315484 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.315506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.315521 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.340997 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.358402 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.375844 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.396930 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.419123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.419171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.419185 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.419205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.419217 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.422718 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.437571 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.464040 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:55Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.520834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.520892 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.520905 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.520921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.520931 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.626264 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.626317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.626330 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.626351 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.626364 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.729674 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.729748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.729757 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.729774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.729786 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.832560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.832640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.832664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.832716 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.832737 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.935293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.935356 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.935369 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.935393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:55 crc kubenswrapper[4758]: I1203 16:56:55.935408 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:55Z","lastTransitionTime":"2025-12-03T16:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.037669 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.037740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.037753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.037770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.037780 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.113496 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:56 crc kubenswrapper[4758]: E1203 16:56:56.113941 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.114200 4758 scope.go:117] "RemoveContainer" containerID="5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.140768 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.140813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.140823 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.140845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.140865 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.244773 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.244862 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.244887 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.244934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.244974 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.347745 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.347794 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.347805 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.347821 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.347834 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.450224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.450267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.450278 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.450299 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.450314 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.521423 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/2.log" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.527890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.528825 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.551824 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.552931 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.552986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.553001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.553026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.553041 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.568146 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.588451 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.606367 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.622821 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.637251 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.648066 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.655732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.655766 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.655777 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.655793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.655804 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.664488 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.681799 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.698275 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.712157 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.728916 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.741160 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.758604 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.758654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.758666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.758703 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.758716 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.761818 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.777582 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.792071 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.807912 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.822389 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:56Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.861235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.861286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.861299 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.861321 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.861331 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.964378 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.964439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.964450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.964467 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:56 crc kubenswrapper[4758]: I1203 16:56:56.964476 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:56Z","lastTransitionTime":"2025-12-03T16:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.067949 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.068032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.068055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.068085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.068104 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.114195 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.114235 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.114389 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:57 crc kubenswrapper[4758]: E1203 16:56:57.114601 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:57 crc kubenswrapper[4758]: E1203 16:56:57.114799 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:57 crc kubenswrapper[4758]: E1203 16:56:57.114861 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.171916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.171981 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.172002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.172030 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.172050 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.274813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.274884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.274896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.274913 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.274924 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.377898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.377989 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.378014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.378051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.378078 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.481955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.481998 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.482008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.482026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.482038 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.534745 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/3.log" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.535490 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/2.log" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.539211 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" exitCode=1 Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.539262 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.539315 4758 scope.go:117] "RemoveContainer" containerID="5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.540293 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 16:56:57 crc kubenswrapper[4758]: E1203 16:56:57.540501 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.562028 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.582070 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.585595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.585941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.586196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.586436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.586725 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.601128 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.620996 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.636786 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.664900 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.682082 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.690160 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.690232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.690242 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.690259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.690270 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.697872 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.712308 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.737356 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5946b52e9b9d0c46d84d8da93661f41765abe6a4f5bad4e43b65f4073bd72955\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:29Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:56:29.138539 6392 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138609 6392 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138675 6392 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.138843 6392 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139142 6392 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139332 6392 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:56:29.139793 6392 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 16:56:29.139821 6392 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:56:29.139841 6392 factory.go:656] Stopping watch factory\\\\nI1203 16:56:29.139858 6392 ovnkube.go:599] Stopped ovnkube\\\\nI1203 16:56:29.139866 6392 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:57Z\\\",\\\"message\\\":\\\":[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 16:56:56.951152 6762 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1203 16:56:56.951163 6762 services_controller.go:453] Built service openshift-kube-storage-version-migrator-operator/metrics template LB for network=default: []services.LB{}\\\\nI1203 16:56:56.951173 6762 services_controller.go:454] Service openshift-kube-storage-version-migrator-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1203 16:56:56.951190 6762 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"7f9b8f25-db1a-4d02-a423-9afc5c2fb83c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.750657 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.765924 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.783153 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.793715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.793792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.793809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.793852 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.793867 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.797875 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.811192 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.822858 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.839552 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.855696 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:57Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.896942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.896983 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.896992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.897008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.897018 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.999397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.999448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.999458 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.999474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:57 crc kubenswrapper[4758]: I1203 16:56:57.999484 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:57Z","lastTransitionTime":"2025-12-03T16:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.101448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.101496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.101507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.101522 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.101533 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.114068 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:56:58 crc kubenswrapper[4758]: E1203 16:56:58.114196 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.124740 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.209515 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.209624 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.209647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.209707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.209739 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.313226 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.313275 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.313285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.313302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.313313 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.416566 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.416628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.416641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.416662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.416693 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.519038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.519073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.519081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.519095 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.519105 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.544805 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/3.log" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.550844 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 16:56:58 crc kubenswrapper[4758]: E1203 16:56:58.551041 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.568958 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.581313 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.595663 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.606580 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.623233 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.623266 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.623276 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.623294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.623306 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.625635 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:57Z\\\",\\\"message\\\":\\\":[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 16:56:56.951152 6762 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1203 16:56:56.951163 6762 services_controller.go:453] Built service openshift-kube-storage-version-migrator-operator/metrics template LB for network=default: []services.LB{}\\\\nI1203 16:56:56.951173 6762 services_controller.go:454] Service openshift-kube-storage-version-migrator-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1203 16:56:56.951190 6762 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"7f9b8f25-db1a-4d02-a423-9afc5c2fb83c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.638498 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.652712 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.667935 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.679461 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.691275 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.702424 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.716897 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.725107 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.725175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.725194 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.725220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.725237 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.730234 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.740872 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.753253 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.766522 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b9d0381-303f-49ba-b044-66ecbb501cf9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://567a6727629cd00ae7f791b31551abbc9fad1a640cff7e997ced687bb5e8bca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffed3a292f0b121961ed8dee1c22d1b70ca5fd6594c46f218a21027f157a403e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffed3a292f0b121961ed8dee1c22d1b70ca5fd6594c46f218a21027f157a403e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.779730 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.793870 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.806712 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:56:58Z is after 2025-08-24T17:21:41Z" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.828180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.828238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.828254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.828277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.828289 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.931064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.931100 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.931108 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.931125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:58 crc kubenswrapper[4758]: I1203 16:56:58.931135 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:58Z","lastTransitionTime":"2025-12-03T16:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.034293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.034340 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.034349 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.034366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.034378 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.113356 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.113418 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:56:59 crc kubenswrapper[4758]: E1203 16:56:59.113512 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:56:59 crc kubenswrapper[4758]: E1203 16:56:59.113591 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.113842 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:56:59 crc kubenswrapper[4758]: E1203 16:56:59.113933 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.137142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.137206 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.137220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.137514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.137531 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.240392 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.240434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.240456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.240559 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.240570 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.344513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.344990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.345061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.345130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.345202 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.448089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.448152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.448172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.448199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.448217 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.551162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.551226 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.551237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.551254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.551265 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.653607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.653652 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.653664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.653702 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.653715 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.757142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.757199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.757214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.757232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.757244 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.859771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.859808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.859819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.859833 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.859843 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.962149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.962179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.962188 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.962203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:56:59 crc kubenswrapper[4758]: I1203 16:56:59.962212 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:56:59Z","lastTransitionTime":"2025-12-03T16:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.064987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.065045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.065059 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.065079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.065094 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.113700 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:00 crc kubenswrapper[4758]: E1203 16:57:00.113996 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.167940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.168008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.168030 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.168064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.168093 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.270518 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.270563 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.270572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.270587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.270597 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.372802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.372838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.372847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.372861 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.372871 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.475144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.475191 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.475203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.475220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.475231 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.577260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.577304 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.577313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.577329 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.577337 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.680086 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.680127 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.680142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.680164 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.680181 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.783045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.783115 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.783132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.783157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.783178 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.885906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.885975 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.885986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.886003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.886013 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.987995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.988028 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.988039 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.988054 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:00 crc kubenswrapper[4758]: I1203 16:57:00.988066 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:00Z","lastTransitionTime":"2025-12-03T16:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.090305 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.090595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.090606 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.090625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.090636 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.113325 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.113359 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.113334 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:01 crc kubenswrapper[4758]: E1203 16:57:01.113469 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:01 crc kubenswrapper[4758]: E1203 16:57:01.113544 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:01 crc kubenswrapper[4758]: E1203 16:57:01.113629 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.192858 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.192953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.192966 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.192985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.192997 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.295841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.295891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.295904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.295923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.295935 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.398966 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.398994 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.399002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.399015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.399023 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.502213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.502253 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.502264 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.502281 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.502291 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.605057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.605147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.605159 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.605176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.605189 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.707832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.707883 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.707895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.707912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.707925 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.810336 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.810389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.810402 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.810420 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.810435 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.912577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.912638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.912648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.912661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:01 crc kubenswrapper[4758]: I1203 16:57:01.912670 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:01Z","lastTransitionTime":"2025-12-03T16:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.016045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.016248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.016357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.016383 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.016413 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.113885 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:02 crc kubenswrapper[4758]: E1203 16:57:02.114028 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.119502 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.119539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.119589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.119613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.119625 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.221660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.221724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.221736 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.221753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.221765 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.324643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.324703 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.324720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.324737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.324752 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.427113 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.427156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.427177 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.427199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.427210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.529632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.529741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.529766 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.529795 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.529815 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.632421 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.632457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.632469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.632484 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.632495 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.735262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.735295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.735304 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.735319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.735328 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.837910 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.838011 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.838026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.838047 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.838060 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.941075 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.941115 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.941128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.941144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:02 crc kubenswrapper[4758]: I1203 16:57:02.941153 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:02Z","lastTransitionTime":"2025-12-03T16:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.044495 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.044553 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.044565 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.044582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.044592 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.114271 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.114355 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:03 crc kubenswrapper[4758]: E1203 16:57:03.114417 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.114460 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:03 crc kubenswrapper[4758]: E1203 16:57:03.114498 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:03 crc kubenswrapper[4758]: E1203 16:57:03.114630 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.146828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.147081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.147237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.147316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.147371 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.250096 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.250143 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.250155 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.250170 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.250182 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.352763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.352824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.352842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.352868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.352885 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.455458 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.455496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.455506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.455521 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.455533 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.558089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.558150 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.558167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.558192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.558210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.661163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.661205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.661218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.661236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.661249 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.763298 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.763335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.763353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.763370 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.763381 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.867329 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.867386 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.867406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.867468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.867486 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.971013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.971062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.971075 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.971097 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:03 crc kubenswrapper[4758]: I1203 16:57:03.971112 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:03Z","lastTransitionTime":"2025-12-03T16:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.074154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.074206 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.074217 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.074236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.074247 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.114366 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:04 crc kubenswrapper[4758]: E1203 16:57:04.114580 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.177547 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.177609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.177630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.177657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.177702 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.280464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.280520 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.280533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.280553 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.280566 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.383053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.383285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.383347 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.383414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.383544 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.486237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.486292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.486304 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.486321 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.486330 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.588416 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.588463 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.588487 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.588509 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.588527 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.691880 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.691929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.691946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.691969 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.691985 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.794227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.794286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.794299 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.794318 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.794328 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.896543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.896578 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.896587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.896605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.896614 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.998359 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.998387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.998395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.998408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:04 crc kubenswrapper[4758]: I1203 16:57:04.998418 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:04Z","lastTransitionTime":"2025-12-03T16:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.100962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.101001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.101044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.101064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.101077 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.113909 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.114019 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.114061 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.114207 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.114281 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.114204 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.126759 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69900819-66e0-4891-bd32-cd06492eea18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:56:02Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 16:55:57.538752 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 16:55:57.543780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2171347958/tls.crt::/tmp/serving-cert-2171347958/tls.key\\\\\\\"\\\\nI1203 16:56:02.764532 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 16:56:02.767163 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 16:56:02.767189 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 16:56:02.767229 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 16:56:02.767240 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 16:56:02.775306 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 16:56:02.776755 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776792 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 16:56:02.776798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 16:56:02.776802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 16:56:02.776806 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 16:56:02.776810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 16:56:02.775804 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 16:56:02.777830 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.138473 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://967df9ddecca91b39811d4429c513ec4e78493241d9f635edeaf643c74df0db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd56c6fd086b63e6dc26cf2f84c6ec40b715b04a427969a3cba132600af3d044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.151219 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e7d6778e28aabc0c42a0fdc17039fcd5bf515956509ae636ad62f2e310b187e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.162867 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ab0180119a91d16c2aa7e434eae53ee89981527b7d4bd93c70b7efbd7bfe480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rl84d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fdssg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.173300 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c4hd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824d1181-466f-47b4-83d7-cc7b79868ff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f334fd6fd3bf8f395f52295151d03ea5fb3c238f0cbbafa9597787b2a684fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-29bhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c4hd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.183408 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b9d0381-303f-49ba-b044-66ecbb501cf9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://567a6727629cd00ae7f791b31551abbc9fad1a640cff7e997ced687bb5e8bca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffed3a292f0b121961ed8dee1c22d1b70ca5fd6594c46f218a21027f157a403e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffed3a292f0b121961ed8dee1c22d1b70ca5fd6594c46f218a21027f157a403e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.195101 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07bf1eee-17d5-4c8a-a214-fca90387fab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16a03835fbaff54452de924f20d40cd8792e0fa259c85599a796cd67cef9564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8161b2663c548700855f6938c19380d401396e743e1985d5b272be44ab56b55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://006251473d463d28ba1876c79865376fa026d6a143cd5a2a6926a1cd46b704e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9663999cd2ff9f269126c80b1fa42725962846fcec937350f5e3b507115cae36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.203736 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.203789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.203804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.203825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.203842 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.211561 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.224779 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6hk5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dc871d6-da79-41c8-b479-0a0d193d4c78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://493d1eb45a511d1897e816db1104fda595ea288fda5db9a06b424bcdf3dc180c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftht5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6hk5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.260473 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c669fa5f-c39f-44f4-ace4-5d5c3cf8267c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6632919627b3bf62553c06aff3331b7fa6ed30fd7147fb409adda38a7a0e17d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518ce53a53b61505dc76f86045c7fcc3edd19666c92a16e9aac9621384680d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://409aebae8b4203b8680c0bd1997323f1bbb7c48ac543def7cd6c072412e002d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://696c87dc62228019f702096af3f6bfa9f0f0327360cea891c71d4cb671bfc426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd8ea88407a069b3cd18e9e3301c7c1119bd75fde562e66999ad15f4cb4e811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ce132b55be17229bbf2611014bd6e4026bbac3c6c9274a580163a297fe8fb86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49fab8e153d3e6208c2a6d7b88bea76f3d9526392fbc3226f054fba135ad3062\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e6cf69a7aa216ec8ea0433387cc893ed71b1fee65008cea415e86e28f5aea5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.285277 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.298719 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zjq67" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06439031-4230-488c-9422-97448eeac39d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbb13886e3db602bfe6c0e3ed1427e0f73bcf4f88d618574a9af89ca586611f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28e921e6ef7c2690e8b9a46a88452db1d18de29c3fedf4b54e705fadcf3caa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://531ac286b6591bd602ebab332c0ea807c16e450874a7ddfccee37eb325edf25f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d43b560d02308031ffa895f807deeae39ed7a87abf0c3ea0712a9d13160a3cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4441d7d5cb3674f2f88b3ade2ad158cd811716551d76609c2040bbf07254c50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d585310cb20ad9446e3a0e7d8fe18843b53006550582f75674c02d15f727e68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bee362b7c1999768f1da34560cace4e4ea8d663a9bf43bcee213a0096736239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kjpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zjq67\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.305886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.305923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.305941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.305955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.305965 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.309915 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-55dnk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgrhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-55dnk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.319157 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afe854a4-7a25-4502-8c41-02b637d14c1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6703e2c6b07eee5bb5504586fef9a8e4166c5e1bb3eac4ad0d7e8f856ca65909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://154fda378a7268ec935ab5090703c956106ef848902f6a6ff8279ebbc9b34c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pvqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gkj9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.328500 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28a94318-e637-479a-9058-86cac6824aa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bef3a6bce639e73e5d4a31f055b1af684bca152cde755fdea8ad9a0e19504e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cf8b284488eae8744ea4c8283e39c7afa846494c8a8fa869b2262d2fd7dc67c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://112792e7200156bd1d9b9165e7a332ef4e19aa5a2f54c252dfada00759a5bc83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:55:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.338504 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad68062a3c0009ad50c37689ab323623fefd2ad573079c37ad005c2d8e27b0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.348285 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.358468 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zwcmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b5429d1-1fc3-4603-93ba-b57b33c2f585\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:50Z\\\",\\\"message\\\":\\\"2025-12-03T16:56:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76\\\\n2025-12-03T16:56:05+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_30b12020-510f-4c36-a24e-2616b968ba76 to /host/opt/cni/bin/\\\\n2025-12-03T16:56:05Z [verbose] multus-daemon started\\\\n2025-12-03T16:56:05Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:56:50Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgt44\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zwcmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.374495 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b40dc3c2-0bc6-447b-ae72-d71650d53f58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:56:57Z\\\",\\\"message\\\":\\\":[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 16:56:56.951152 6762 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1203 16:56:56.951163 6762 services_controller.go:453] Built service openshift-kube-storage-version-migrator-operator/metrics template LB for network=default: []services.LB{}\\\\nI1203 16:56:56.951173 6762 services_controller.go:454] Service openshift-kube-storage-version-migrator-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1203 16:56:56.951190 6762 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"7f9b8f25-db1a-4d02-a423-9afc5c2fb83c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:56:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:56:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:56:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47xn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:56:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9p24\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.408260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.408286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.408294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.408309 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.408318 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.475528 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.475592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.475609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.475635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.475652 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.498653 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.504723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.504772 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.504782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.504799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.504812 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.522365 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.526128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.526178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.526196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.526220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.526238 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.547118 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.551511 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.551539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.551551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.551568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.551581 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.569102 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.573617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.573658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.573667 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.573698 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.573711 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.589141 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8c04589e-31a6-4e69-8c1b-e1795a06a3f9\\\",\\\"systemUUID\\\":\\\"8cc52a80-b3b7-4ab3-8e60-5535753f3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:57:05 crc kubenswrapper[4758]: E1203 16:57:05.589265 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.590612 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.590639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.590648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.590663 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.590673 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.693665 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.693755 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.693773 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.693809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.693828 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.796573 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.796629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.796644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.796661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.796674 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.899269 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.899310 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.899321 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.899336 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:05 crc kubenswrapper[4758]: I1203 16:57:05.899346 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:05Z","lastTransitionTime":"2025-12-03T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.001606 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.001638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.001647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.001660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.001669 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.104428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.104504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.104552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.104573 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.104585 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.113919 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:06 crc kubenswrapper[4758]: E1203 16:57:06.114071 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.207381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.207432 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.207443 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.207461 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.207471 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.310072 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.310109 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.310118 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.310133 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.310142 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.412290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.412340 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.412350 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.412364 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.412373 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.515419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.515460 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.515470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.515486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.515496 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.618153 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.618200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.618215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.618234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.618245 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.720587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.720625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.720635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.720652 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.720662 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.823938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.824003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.824022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.824045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.824062 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.927438 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.927486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.927498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.927519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:06 crc kubenswrapper[4758]: I1203 16:57:06.927532 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:06Z","lastTransitionTime":"2025-12-03T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.015721 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.015981 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:11.015943056 +0000 UTC m=+146.217319957 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.016358 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.016409 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.016519 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.016535 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.016610 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:58:11.016590385 +0000 UTC m=+146.217967276 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.016638 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:58:11.016626986 +0000 UTC m=+146.218003877 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.030457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.030486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.030497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.030512 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.030523 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.113856 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.114003 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.114215 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.114221 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.114324 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.114388 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.117249 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.117474 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117543 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117601 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117612 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117625 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117637 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117654 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117734 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:58:11.117713117 +0000 UTC m=+146.319090169 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:57:07 crc kubenswrapper[4758]: E1203 16:57:07.117794 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:58:11.117745778 +0000 UTC m=+146.319122729 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.134946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.135026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.135049 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.135085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.135106 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.237728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.237814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.237842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.237878 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.237904 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.341157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.341207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.341218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.341236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.341252 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.444608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.444659 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.444671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.444706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.444715 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.547913 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.547964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.547979 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.547999 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.548012 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.651753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.651813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.651829 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.651857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.651875 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.756343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.756449 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.756469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.756536 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.756557 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.861071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.861125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.861138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.861157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.861170 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.964029 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.964080 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.964091 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.964111 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:07 crc kubenswrapper[4758]: I1203 16:57:07.964172 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:07Z","lastTransitionTime":"2025-12-03T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.067285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.067333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.067342 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.067361 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.067374 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.113395 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:08 crc kubenswrapper[4758]: E1203 16:57:08.113659 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.170271 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.170302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.170311 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.170327 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.170336 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.273002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.273031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.273040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.273058 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.273075 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.382111 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.382200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.382222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.382490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.382739 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.485982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.486047 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.486056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.486068 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.486078 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.587956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.587999 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.588011 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.588027 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.588062 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.690701 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.690738 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.690747 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.690762 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.690772 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.793874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.793921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.793932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.793947 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.793956 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.896031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.896074 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.896089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.896105 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.896115 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.997935 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.997972 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.997982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.997998 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:08 crc kubenswrapper[4758]: I1203 16:57:08.998009 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:08Z","lastTransitionTime":"2025-12-03T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.100541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.100595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.100611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.100636 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.100653 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.114211 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.114336 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:09 crc kubenswrapper[4758]: E1203 16:57:09.114373 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:09 crc kubenswrapper[4758]: E1203 16:57:09.114488 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.114316 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:09 crc kubenswrapper[4758]: E1203 16:57:09.114606 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.203953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.203992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.204002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.204015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.204025 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.307079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.307136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.307150 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.307171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.307185 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.410792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.410860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.410875 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.410897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.410915 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.514333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.514389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.514403 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.514424 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.514442 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.617636 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.617735 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.617752 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.617778 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.617795 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.720968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.721064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.721083 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.721108 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.721124 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.824769 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.824868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.824886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.824914 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.824936 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.927300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.927348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.927357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.927403 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:09 crc kubenswrapper[4758]: I1203 16:57:09.927422 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:09Z","lastTransitionTime":"2025-12-03T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.030569 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.030630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.030644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.030667 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.030699 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.114328 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:10 crc kubenswrapper[4758]: E1203 16:57:10.114569 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.115845 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 16:57:10 crc kubenswrapper[4758]: E1203 16:57:10.116164 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.133983 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.134058 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.134098 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.134136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.134162 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.238260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.238353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.238382 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.238416 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.238442 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.341480 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.341570 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.341590 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.341620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.341641 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.444510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.444601 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.444619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.444642 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.444656 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.547638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.547744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.547770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.547799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.547820 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.650610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.650738 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.650759 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.650786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.650804 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.754172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.754222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.754232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.754249 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.754261 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.856995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.857038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.857049 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.857247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.857265 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.960348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.960385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.960396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.960413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:10 crc kubenswrapper[4758]: I1203 16:57:10.960425 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:10Z","lastTransitionTime":"2025-12-03T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.063225 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.063644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.063762 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.063869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.063950 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.114225 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.114264 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.114226 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:11 crc kubenswrapper[4758]: E1203 16:57:11.114400 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:11 crc kubenswrapper[4758]: E1203 16:57:11.114770 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:11 crc kubenswrapper[4758]: E1203 16:57:11.114881 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.166473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.166516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.166533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.166549 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.166560 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.269866 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.269921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.269943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.269999 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.270018 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.373620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.373813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.373877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.373912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.373974 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.477517 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.477583 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.477600 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.477627 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.477649 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.581172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.581290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.581320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.581354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.581379 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.684216 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.684253 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.684264 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.684280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.684290 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.786882 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.786945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.786963 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.786986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.787004 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.890771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.890857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.890870 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.890890 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.890902 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.994068 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.994124 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.994138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.994192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:11 crc kubenswrapper[4758]: I1203 16:57:11.994208 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:11Z","lastTransitionTime":"2025-12-03T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.097918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.098035 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.098060 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.098097 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.098121 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.114075 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:12 crc kubenswrapper[4758]: E1203 16:57:12.114216 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.202135 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.202226 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.202246 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.202282 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.202303 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.305713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.306165 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.306261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.306373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.306460 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.410025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.410418 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.410536 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.410666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.410815 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.514373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.515170 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.515260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.515429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.515663 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.618821 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.618888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.618937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.618968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.618994 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.721943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.722011 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.722025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.722047 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.722064 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.824640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.824848 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.824948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.825058 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.825075 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.927750 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.927813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.927823 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.927838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:12 crc kubenswrapper[4758]: I1203 16:57:12.927850 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:12Z","lastTransitionTime":"2025-12-03T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.030758 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.030806 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.030816 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.030834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.030844 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.113611 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.113638 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.113743 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:13 crc kubenswrapper[4758]: E1203 16:57:13.114493 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:13 crc kubenswrapper[4758]: E1203 16:57:13.114556 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:13 crc kubenswrapper[4758]: E1203 16:57:13.114732 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.138541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.138618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.138634 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.138655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.138670 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.241527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.241601 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.241620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.241649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.241668 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.344306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.344369 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.344381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.344401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.344412 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.447572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.447625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.447637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.447654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.447666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.551723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.551780 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.551789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.551808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.551819 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.655423 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.655496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.655510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.655534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.655553 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.758772 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.758879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.758923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.758945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.758962 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.861009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.861056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.861067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.861089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.861103 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.964801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.964854 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.964867 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.964888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:13 crc kubenswrapper[4758]: I1203 16:57:13.964902 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:13Z","lastTransitionTime":"2025-12-03T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.067550 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.067613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.067630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.067658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.067700 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.113363 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:14 crc kubenswrapper[4758]: E1203 16:57:14.113558 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.171109 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.171162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.171174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.171193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.171210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.274001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.274050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.274066 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.274085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.274097 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.377789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.377855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.377868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.377884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.377895 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.480399 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.480444 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.480452 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.480470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.480481 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.584087 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.584168 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.584193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.584231 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.584260 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.687535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.687579 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.687589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.687608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.687625 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.790228 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.790294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.790306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.790329 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.790343 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.892485 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.892551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.892564 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.892590 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.892603 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.996036 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.996103 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.996123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.996151 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:14 crc kubenswrapper[4758]: I1203 16:57:14.996169 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:14Z","lastTransitionTime":"2025-12-03T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.101247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.101883 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.102411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.103138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.103397 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.113985 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.114401 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.114657 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:15 crc kubenswrapper[4758]: E1203 16:57:15.115030 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:15 crc kubenswrapper[4758]: E1203 16:57:15.115283 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:15 crc kubenswrapper[4758]: E1203 16:57:15.115837 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.196420 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-zwcmr" podStartSLOduration=72.196397883 podStartE2EDuration="1m12.196397883s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.165314411 +0000 UTC m=+90.366691302" watchObservedRunningTime="2025-12-03 16:57:15.196397883 +0000 UTC m=+90.397774764" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.206073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.206113 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.206123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.206139 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.206151 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.232530 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gkj9l" podStartSLOduration=71.232480219 podStartE2EDuration="1m11.232480219s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.230863622 +0000 UTC m=+90.432240503" watchObservedRunningTime="2025-12-03 16:57:15.232480219 +0000 UTC m=+90.433857090" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.273862 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.273841716 podStartE2EDuration="1m9.273841716s" podCreationTimestamp="2025-12-03 16:56:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.254185002 +0000 UTC m=+90.455561863" watchObservedRunningTime="2025-12-03 16:57:15.273841716 +0000 UTC m=+90.475218577" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.305636 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podStartSLOduration=72.305615119 podStartE2EDuration="1m12.305615119s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.296440455 +0000 UTC m=+90.497817316" watchObservedRunningTime="2025-12-03 16:57:15.305615119 +0000 UTC m=+90.506991980" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.306447 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-c4hd9" podStartSLOduration=72.306441712 podStartE2EDuration="1m12.306441712s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.306418981 +0000 UTC m=+90.507795852" watchObservedRunningTime="2025-12-03 16:57:15.306441712 +0000 UTC m=+90.507818563" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.308954 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.308998 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.309008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.309026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.309036 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.344535 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.344502815 podStartE2EDuration="1m12.344502815s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.329073232 +0000 UTC m=+90.530450113" watchObservedRunningTime="2025-12-03 16:57:15.344502815 +0000 UTC m=+90.545879676" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.372000 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6hk5h" podStartSLOduration=72.371981083 podStartE2EDuration="1m12.371981083s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.371786338 +0000 UTC m=+90.573163199" watchObservedRunningTime="2025-12-03 16:57:15.371981083 +0000 UTC m=+90.573357944" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.399176 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.399155584 podStartE2EDuration="35.399155584s" podCreationTimestamp="2025-12-03 16:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.398235267 +0000 UTC m=+90.599612128" watchObservedRunningTime="2025-12-03 16:57:15.399155584 +0000 UTC m=+90.600532455" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.399448 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.399442502 podStartE2EDuration="17.399442502s" podCreationTimestamp="2025-12-03 16:56:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.382436704 +0000 UTC m=+90.583813585" watchObservedRunningTime="2025-12-03 16:57:15.399442502 +0000 UTC m=+90.600819363" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.411448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.411492 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.411502 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.411517 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.411529 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.416077 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zjq67" podStartSLOduration=72.416067599 podStartE2EDuration="1m12.416067599s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.415538284 +0000 UTC m=+90.616915145" watchObservedRunningTime="2025-12-03 16:57:15.416067599 +0000 UTC m=+90.617444460" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.455262 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=69.455244174 podStartE2EDuration="1m9.455244174s" podCreationTimestamp="2025-12-03 16:56:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:15.454297387 +0000 UTC m=+90.655674238" watchObservedRunningTime="2025-12-03 16:57:15.455244174 +0000 UTC m=+90.656621035" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.513896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.514169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.514235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.514304 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.514383 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.616980 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.617020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.617057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.617075 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.617085 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.720309 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.720413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.720429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.720448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.720461 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.823647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.823754 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.823788 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.823809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.823828 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.872313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.872366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.872378 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.872395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.872405 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:57:15Z","lastTransitionTime":"2025-12-03T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.925704 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86"] Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.926136 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.928021 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.928948 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.930465 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 16:57:15 crc kubenswrapper[4758]: I1203 16:57:15.930791 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.011948 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/31895278-b881-49ce-b331-2caa4158a3e1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.012039 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/31895278-b881-49ce-b331-2caa4158a3e1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.012068 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31895278-b881-49ce-b331-2caa4158a3e1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.012093 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31895278-b881-49ce-b331-2caa4158a3e1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.012119 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/31895278-b881-49ce-b331-2caa4158a3e1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113312 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/31895278-b881-49ce-b331-2caa4158a3e1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113381 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/31895278-b881-49ce-b331-2caa4158a3e1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/31895278-b881-49ce-b331-2caa4158a3e1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113445 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31895278-b881-49ce-b331-2caa4158a3e1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113461 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31895278-b881-49ce-b331-2caa4158a3e1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113577 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/31895278-b881-49ce-b331-2caa4158a3e1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113583 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.113579 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/31895278-b881-49ce-b331-2caa4158a3e1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: E1203 16:57:16.113788 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.114487 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/31895278-b881-49ce-b331-2caa4158a3e1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.120743 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31895278-b881-49ce-b331-2caa4158a3e1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.140272 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31895278-b881-49ce-b331-2caa4158a3e1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7jg86\" (UID: \"31895278-b881-49ce-b331-2caa4158a3e1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.241543 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" Dec 03 16:57:16 crc kubenswrapper[4758]: W1203 16:57:16.255783 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31895278_b881_49ce_b331_2caa4158a3e1.slice/crio-2b40376b0d1eaddb1049174dcc94a4252cd8084573401b375300e98dca8872b4 WatchSource:0}: Error finding container 2b40376b0d1eaddb1049174dcc94a4252cd8084573401b375300e98dca8872b4: Status 404 returned error can't find the container with id 2b40376b0d1eaddb1049174dcc94a4252cd8084573401b375300e98dca8872b4 Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.611934 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" event={"ID":"31895278-b881-49ce-b331-2caa4158a3e1","Type":"ContainerStarted","Data":"9ccf15a0f77e002d4fa7044613c0bb3fd5bf1bd99a3cd41e3021f4de25e68869"} Dec 03 16:57:16 crc kubenswrapper[4758]: I1203 16:57:16.611981 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" event={"ID":"31895278-b881-49ce-b331-2caa4158a3e1","Type":"ContainerStarted","Data":"2b40376b0d1eaddb1049174dcc94a4252cd8084573401b375300e98dca8872b4"} Dec 03 16:57:17 crc kubenswrapper[4758]: I1203 16:57:17.114345 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:17 crc kubenswrapper[4758]: I1203 16:57:17.114391 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:17 crc kubenswrapper[4758]: I1203 16:57:17.114457 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:17 crc kubenswrapper[4758]: E1203 16:57:17.114594 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:17 crc kubenswrapper[4758]: E1203 16:57:17.114850 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:17 crc kubenswrapper[4758]: E1203 16:57:17.115058 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:18 crc kubenswrapper[4758]: I1203 16:57:18.114220 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:18 crc kubenswrapper[4758]: E1203 16:57:18.114413 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:19 crc kubenswrapper[4758]: I1203 16:57:19.114325 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:19 crc kubenswrapper[4758]: I1203 16:57:19.114428 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:19 crc kubenswrapper[4758]: E1203 16:57:19.114552 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:19 crc kubenswrapper[4758]: I1203 16:57:19.114493 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:19 crc kubenswrapper[4758]: E1203 16:57:19.114823 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:19 crc kubenswrapper[4758]: E1203 16:57:19.114976 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:20 crc kubenswrapper[4758]: I1203 16:57:20.113334 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:20 crc kubenswrapper[4758]: E1203 16:57:20.113553 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:21 crc kubenswrapper[4758]: I1203 16:57:21.114035 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:21 crc kubenswrapper[4758]: I1203 16:57:21.114146 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:21 crc kubenswrapper[4758]: I1203 16:57:21.114226 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:21 crc kubenswrapper[4758]: E1203 16:57:21.114417 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:21 crc kubenswrapper[4758]: E1203 16:57:21.114599 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:21 crc kubenswrapper[4758]: E1203 16:57:21.114994 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:22 crc kubenswrapper[4758]: I1203 16:57:22.114059 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:22 crc kubenswrapper[4758]: E1203 16:57:22.114236 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:22 crc kubenswrapper[4758]: I1203 16:57:22.272227 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:22 crc kubenswrapper[4758]: E1203 16:57:22.272495 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:57:22 crc kubenswrapper[4758]: E1203 16:57:22.272663 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs podName:5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af nodeName:}" failed. No retries permitted until 2025-12-03 16:58:26.272622433 +0000 UTC m=+161.473999454 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs") pod "network-metrics-daemon-55dnk" (UID: "5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:57:23 crc kubenswrapper[4758]: I1203 16:57:23.113662 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:23 crc kubenswrapper[4758]: I1203 16:57:23.113662 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:23 crc kubenswrapper[4758]: E1203 16:57:23.113832 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:23 crc kubenswrapper[4758]: I1203 16:57:23.114261 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:23 crc kubenswrapper[4758]: E1203 16:57:23.114486 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:23 crc kubenswrapper[4758]: E1203 16:57:23.114632 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:23 crc kubenswrapper[4758]: I1203 16:57:23.115057 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 16:57:23 crc kubenswrapper[4758]: E1203 16:57:23.115363 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:57:24 crc kubenswrapper[4758]: I1203 16:57:24.113660 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:24 crc kubenswrapper[4758]: E1203 16:57:24.114053 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:25 crc kubenswrapper[4758]: I1203 16:57:25.114095 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:25 crc kubenswrapper[4758]: I1203 16:57:25.114202 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:25 crc kubenswrapper[4758]: I1203 16:57:25.114249 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:25 crc kubenswrapper[4758]: E1203 16:57:25.115972 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:25 crc kubenswrapper[4758]: E1203 16:57:25.116060 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:25 crc kubenswrapper[4758]: E1203 16:57:25.116145 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:26 crc kubenswrapper[4758]: I1203 16:57:26.114431 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:26 crc kubenswrapper[4758]: E1203 16:57:26.114988 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:27 crc kubenswrapper[4758]: I1203 16:57:27.114406 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:27 crc kubenswrapper[4758]: I1203 16:57:27.114509 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:27 crc kubenswrapper[4758]: I1203 16:57:27.114654 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:27 crc kubenswrapper[4758]: E1203 16:57:27.114901 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:27 crc kubenswrapper[4758]: E1203 16:57:27.115621 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:27 crc kubenswrapper[4758]: E1203 16:57:27.115737 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:28 crc kubenswrapper[4758]: I1203 16:57:28.113730 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:28 crc kubenswrapper[4758]: E1203 16:57:28.113918 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:29 crc kubenswrapper[4758]: I1203 16:57:29.113583 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:29 crc kubenswrapper[4758]: I1203 16:57:29.113622 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:29 crc kubenswrapper[4758]: I1203 16:57:29.113976 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:29 crc kubenswrapper[4758]: E1203 16:57:29.114220 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:29 crc kubenswrapper[4758]: E1203 16:57:29.114354 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:29 crc kubenswrapper[4758]: E1203 16:57:29.114511 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:30 crc kubenswrapper[4758]: I1203 16:57:30.113704 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:30 crc kubenswrapper[4758]: E1203 16:57:30.113884 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:31 crc kubenswrapper[4758]: I1203 16:57:31.114148 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:31 crc kubenswrapper[4758]: E1203 16:57:31.114398 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:31 crc kubenswrapper[4758]: I1203 16:57:31.114181 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:31 crc kubenswrapper[4758]: E1203 16:57:31.114567 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:31 crc kubenswrapper[4758]: I1203 16:57:31.114135 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:31 crc kubenswrapper[4758]: E1203 16:57:31.114753 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:32 crc kubenswrapper[4758]: I1203 16:57:32.113980 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:32 crc kubenswrapper[4758]: E1203 16:57:32.114163 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:33 crc kubenswrapper[4758]: I1203 16:57:33.203534 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:33 crc kubenswrapper[4758]: E1203 16:57:33.204119 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:33 crc kubenswrapper[4758]: I1203 16:57:33.203717 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:33 crc kubenswrapper[4758]: E1203 16:57:33.204202 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:33 crc kubenswrapper[4758]: I1203 16:57:33.203656 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:33 crc kubenswrapper[4758]: E1203 16:57:33.204261 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:34 crc kubenswrapper[4758]: I1203 16:57:34.114338 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:34 crc kubenswrapper[4758]: E1203 16:57:34.115311 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:34 crc kubenswrapper[4758]: I1203 16:57:34.115767 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 16:57:34 crc kubenswrapper[4758]: E1203 16:57:34.116026 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p9p24_openshift-ovn-kubernetes(b40dc3c2-0bc6-447b-ae72-d71650d53f58)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" Dec 03 16:57:35 crc kubenswrapper[4758]: I1203 16:57:35.120237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:35 crc kubenswrapper[4758]: I1203 16:57:35.120237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:35 crc kubenswrapper[4758]: E1203 16:57:35.120449 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:35 crc kubenswrapper[4758]: E1203 16:57:35.120664 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:35 crc kubenswrapper[4758]: I1203 16:57:35.120263 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:35 crc kubenswrapper[4758]: E1203 16:57:35.120912 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:36 crc kubenswrapper[4758]: I1203 16:57:36.113591 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:36 crc kubenswrapper[4758]: E1203 16:57:36.113793 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.113439 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.113439 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.113538 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:37 crc kubenswrapper[4758]: E1203 16:57:37.114023 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:37 crc kubenswrapper[4758]: E1203 16:57:37.114209 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:37 crc kubenswrapper[4758]: E1203 16:57:37.114377 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.736226 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/1.log" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.737016 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/0.log" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.737120 4758 generic.go:334] "Generic (PLEG): container finished" podID="4b5429d1-1fc3-4603-93ba-b57b33c2f585" containerID="949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c" exitCode=1 Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.737179 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerDied","Data":"949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c"} Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.737257 4758 scope.go:117] "RemoveContainer" containerID="3d5741601f8162f44ee6ab00234ed087db02f73a1e23e8be21ab5eb57a1020fb" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.737975 4758 scope.go:117] "RemoveContainer" containerID="949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c" Dec 03 16:57:37 crc kubenswrapper[4758]: E1203 16:57:37.738289 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-zwcmr_openshift-multus(4b5429d1-1fc3-4603-93ba-b57b33c2f585)\"" pod="openshift-multus/multus-zwcmr" podUID="4b5429d1-1fc3-4603-93ba-b57b33c2f585" Dec 03 16:57:37 crc kubenswrapper[4758]: I1203 16:57:37.769239 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7jg86" podStartSLOduration=94.769213999 podStartE2EDuration="1m34.769213999s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:16.639034747 +0000 UTC m=+91.840411608" watchObservedRunningTime="2025-12-03 16:57:37.769213999 +0000 UTC m=+112.970590900" Dec 03 16:57:38 crc kubenswrapper[4758]: I1203 16:57:38.113330 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:38 crc kubenswrapper[4758]: E1203 16:57:38.113523 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:38 crc kubenswrapper[4758]: I1203 16:57:38.744208 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/1.log" Dec 03 16:57:39 crc kubenswrapper[4758]: I1203 16:57:39.114332 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:39 crc kubenswrapper[4758]: I1203 16:57:39.114354 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:39 crc kubenswrapper[4758]: I1203 16:57:39.114548 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:39 crc kubenswrapper[4758]: E1203 16:57:39.114770 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:39 crc kubenswrapper[4758]: E1203 16:57:39.114950 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:39 crc kubenswrapper[4758]: E1203 16:57:39.115185 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:40 crc kubenswrapper[4758]: I1203 16:57:40.114091 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:40 crc kubenswrapper[4758]: E1203 16:57:40.114292 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:41 crc kubenswrapper[4758]: I1203 16:57:41.113483 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:41 crc kubenswrapper[4758]: I1203 16:57:41.113633 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:41 crc kubenswrapper[4758]: E1203 16:57:41.113671 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:41 crc kubenswrapper[4758]: I1203 16:57:41.113740 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:41 crc kubenswrapper[4758]: E1203 16:57:41.113921 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:41 crc kubenswrapper[4758]: E1203 16:57:41.113963 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:42 crc kubenswrapper[4758]: I1203 16:57:42.113471 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:42 crc kubenswrapper[4758]: E1203 16:57:42.113661 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:43 crc kubenswrapper[4758]: I1203 16:57:43.113647 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:43 crc kubenswrapper[4758]: I1203 16:57:43.113825 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:43 crc kubenswrapper[4758]: E1203 16:57:43.114003 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:43 crc kubenswrapper[4758]: I1203 16:57:43.114771 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:43 crc kubenswrapper[4758]: E1203 16:57:43.114980 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:43 crc kubenswrapper[4758]: E1203 16:57:43.115178 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:44 crc kubenswrapper[4758]: I1203 16:57:44.113590 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:44 crc kubenswrapper[4758]: E1203 16:57:44.114116 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:45 crc kubenswrapper[4758]: E1203 16:57:45.072431 4758 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 16:57:45 crc kubenswrapper[4758]: I1203 16:57:45.113456 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:45 crc kubenswrapper[4758]: I1203 16:57:45.113505 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:45 crc kubenswrapper[4758]: I1203 16:57:45.113562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:45 crc kubenswrapper[4758]: E1203 16:57:45.115549 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:45 crc kubenswrapper[4758]: E1203 16:57:45.115728 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:45 crc kubenswrapper[4758]: E1203 16:57:45.115963 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:45 crc kubenswrapper[4758]: E1203 16:57:45.217372 4758 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 16:57:46 crc kubenswrapper[4758]: I1203 16:57:46.113351 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:46 crc kubenswrapper[4758]: E1203 16:57:46.113535 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:46 crc kubenswrapper[4758]: I1203 16:57:46.114655 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 16:57:46 crc kubenswrapper[4758]: I1203 16:57:46.773046 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/3.log" Dec 03 16:57:46 crc kubenswrapper[4758]: I1203 16:57:46.776921 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerStarted","Data":"b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958"} Dec 03 16:57:46 crc kubenswrapper[4758]: I1203 16:57:46.777536 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:57:47 crc kubenswrapper[4758]: I1203 16:57:47.042537 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podStartSLOduration=104.042506317 podStartE2EDuration="1m44.042506317s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:57:46.813451851 +0000 UTC m=+122.014828732" watchObservedRunningTime="2025-12-03 16:57:47.042506317 +0000 UTC m=+122.243883178" Dec 03 16:57:47 crc kubenswrapper[4758]: I1203 16:57:47.042881 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-55dnk"] Dec 03 16:57:47 crc kubenswrapper[4758]: I1203 16:57:47.043028 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:47 crc kubenswrapper[4758]: E1203 16:57:47.043145 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:47 crc kubenswrapper[4758]: I1203 16:57:47.113819 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:47 crc kubenswrapper[4758]: I1203 16:57:47.113872 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:47 crc kubenswrapper[4758]: I1203 16:57:47.113879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:47 crc kubenswrapper[4758]: E1203 16:57:47.114405 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:47 crc kubenswrapper[4758]: E1203 16:57:47.114520 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:47 crc kubenswrapper[4758]: E1203 16:57:47.114238 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.113610 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.113637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.113637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:49 crc kubenswrapper[4758]: E1203 16:57:49.113954 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:49 crc kubenswrapper[4758]: E1203 16:57:49.114156 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.114185 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:49 crc kubenswrapper[4758]: E1203 16:57:49.114269 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.114378 4758 scope.go:117] "RemoveContainer" containerID="949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c" Dec 03 16:57:49 crc kubenswrapper[4758]: E1203 16:57:49.114382 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.791558 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/1.log" Dec 03 16:57:49 crc kubenswrapper[4758]: I1203 16:57:49.791651 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerStarted","Data":"01a2e2844e1250fbdb9021acd40f8f5ec395720ebc71734423c63cdf86bde9fd"} Dec 03 16:57:50 crc kubenswrapper[4758]: E1203 16:57:50.218935 4758 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 16:57:51 crc kubenswrapper[4758]: I1203 16:57:51.113982 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:51 crc kubenswrapper[4758]: I1203 16:57:51.114101 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:51 crc kubenswrapper[4758]: E1203 16:57:51.114343 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:51 crc kubenswrapper[4758]: I1203 16:57:51.114466 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:51 crc kubenswrapper[4758]: E1203 16:57:51.114528 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:51 crc kubenswrapper[4758]: E1203 16:57:51.114753 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:51 crc kubenswrapper[4758]: I1203 16:57:51.114101 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:51 crc kubenswrapper[4758]: E1203 16:57:51.114985 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:53 crc kubenswrapper[4758]: I1203 16:57:53.113882 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:53 crc kubenswrapper[4758]: I1203 16:57:53.113938 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:53 crc kubenswrapper[4758]: I1203 16:57:53.114038 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:53 crc kubenswrapper[4758]: I1203 16:57:53.114054 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:53 crc kubenswrapper[4758]: E1203 16:57:53.114223 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:53 crc kubenswrapper[4758]: E1203 16:57:53.114341 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:53 crc kubenswrapper[4758]: E1203 16:57:53.114473 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:53 crc kubenswrapper[4758]: E1203 16:57:53.114602 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:55 crc kubenswrapper[4758]: I1203 16:57:55.114167 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:55 crc kubenswrapper[4758]: I1203 16:57:55.114218 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:55 crc kubenswrapper[4758]: I1203 16:57:55.114180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:55 crc kubenswrapper[4758]: I1203 16:57:55.115623 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:55 crc kubenswrapper[4758]: E1203 16:57:55.115626 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:57:55 crc kubenswrapper[4758]: E1203 16:57:55.115760 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:57:55 crc kubenswrapper[4758]: E1203 16:57:55.115900 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-55dnk" podUID="5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af" Dec 03 16:57:55 crc kubenswrapper[4758]: E1203 16:57:55.116188 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.610128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.654216 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f4wqg"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.655023 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.668521 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.670117 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.670194 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.670365 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.670357 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7lnzm"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.670537 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.670726 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.671066 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.671266 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.671401 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.671501 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.671719 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.672747 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.673160 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94vfh"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.673750 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.674023 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.674549 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.675488 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.676201 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hqwcv"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.676602 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nqrwv"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.676977 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.677176 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8dmlt"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.677347 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.677712 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.677719 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.677785 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.677844 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.681238 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.682277 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.682435 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.687183 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.687471 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.687749 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.690252 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.690839 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qkmml"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.691320 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.691630 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.691927 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.692117 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.692396 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.703568 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.704615 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.705836 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.705964 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.706049 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.706148 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.706138 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.706251 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.707598 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.720481 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722159 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722227 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722344 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722372 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722394 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4l6l4"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722436 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722471 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722495 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722514 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722527 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722746 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722749 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.722985 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723082 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723128 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qjwkt"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723312 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723137 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723096 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723627 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723192 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723200 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723760 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723262 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723284 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723284 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723381 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.723388 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.724166 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.724302 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.724445 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.725118 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.725307 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.725615 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.725790 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.726146 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.727383 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.727629 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.727788 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.727903 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.728011 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.728132 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.728513 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.728747 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.731938 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.732069 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.732164 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.732465 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.733568 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.733742 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.733857 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.737387 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.740311 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.742752 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.742946 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.743035 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.743151 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.743487 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.743724 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.745515 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.746542 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.746792 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x976j"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.746886 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.747459 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.747484 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kzv6z"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.747627 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.747879 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.748021 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.748200 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.748298 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.748808 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.748879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.748302 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.749264 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.749328 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.749343 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.749526 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.774130 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.774503 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.776663 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.778392 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.777075 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.778024 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.776720 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-etcd-client\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.779195 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-encryption-config\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.779240 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-audit-dir\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805136 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805231 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805311 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-978qs\" (UniqueName: \"kubernetes.io/projected/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-kube-api-access-978qs\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805375 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-etcd-serving-ca\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805426 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-serving-cert\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805461 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-config\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e9ea76f4-50d8-416a-85dd-897f08112fa3-audit-dir\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805501 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805540 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-node-pullsecrets\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805562 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-audit-policies\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805586 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805612 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-etcd-client\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805628 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-audit\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805646 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-image-import-ca\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805659 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-encryption-config\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805696 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8rrt\" (UniqueName: \"kubernetes.io/projected/e9ea76f4-50d8-416a-85dd-897f08112fa3-kube-api-access-k8rrt\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.805718 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-serving-cert\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.806538 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.806789 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.807104 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.807517 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.807692 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4f6xq"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.807931 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.808260 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.807697 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.808266 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.809271 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6r465"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.810003 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.810529 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.810902 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.811568 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.813990 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.812195 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.815144 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.815964 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.816270 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.816295 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.817191 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.817846 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.819242 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.819718 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.821365 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.822382 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nnc2j"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.822742 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.823539 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jxqhr"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.824242 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.825865 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-srplx"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.826044 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.826937 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.827036 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.827830 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.828418 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.828868 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.829551 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f4wqg"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.833662 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7lnzm"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.833735 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.834563 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5hdt9"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.835103 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.835732 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.837931 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.838468 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.839209 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8sqq"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.839493 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.839635 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.840099 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.841072 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.843848 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94vfh"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.843941 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.844094 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.845787 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.846716 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4l6l4"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.847658 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.848998 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4f6xq"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.855850 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hqwcv"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.856727 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qjwkt"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.859222 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nqrwv"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.860250 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.861798 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qkmml"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.864813 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.867329 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kzv6z"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.868490 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.869524 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8dmlt"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.870663 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6r465"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.871706 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.872783 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-c9b57"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.873973 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.874244 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.875417 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.876579 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.877754 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.878904 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-srplx"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.880018 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hvg6p"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.880896 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.881102 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.881894 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.884779 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.889122 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nnc2j"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.893238 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.893464 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x976j"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.893484 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.895649 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.898839 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c9b57"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.900023 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.900738 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.902223 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.903663 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8sqq"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.905286 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5hdt9"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.908504 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.908606 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2gw5x"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.909656 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-dcg47"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.910187 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.949268 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.949953 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-serving-cert\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.949992 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-etcd-client\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950039 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-encryption-config\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950063 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-audit-dir\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950084 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950108 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-978qs\" (UniqueName: \"kubernetes.io/projected/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-kube-api-access-978qs\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950141 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-etcd-serving-ca\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950175 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-serving-cert\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950203 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e9ea76f4-50d8-416a-85dd-897f08112fa3-audit-dir\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950239 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-config\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950262 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-node-pullsecrets\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950351 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-audit-policies\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-etcd-client\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-audit\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950422 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-image-import-ca\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950440 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-encryption-config\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.950459 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8rrt\" (UniqueName: \"kubernetes.io/projected/e9ea76f4-50d8-416a-85dd-897f08112fa3-kube-api-access-k8rrt\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.951629 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.951720 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dcg47"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.951924 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.952633 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-node-pullsecrets\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.953007 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-config\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.953086 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.953185 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.953803 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.953977 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-etcd-serving-ca\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.954030 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e9ea76f4-50d8-416a-85dd-897f08112fa3-audit-policies\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.954100 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e9ea76f4-50d8-416a-85dd-897f08112fa3-audit-dir\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.954758 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-audit\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.954986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-audit-dir\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.955367 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-image-import-ca\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.957499 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2gw5x"] Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.957742 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-encryption-config\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.957801 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-encryption-config\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.957809 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-serving-cert\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.958077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-serving-cert\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.958978 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e9ea76f4-50d8-416a-85dd-897f08112fa3-etcd-client\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.959604 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.964116 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-etcd-client\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.980342 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 16:57:56 crc kubenswrapper[4758]: I1203 16:57:56.999351 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.020152 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.039898 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.059521 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.079653 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.099903 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.113589 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.113965 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.114075 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.114263 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.120456 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.179959 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.200292 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.223739 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.239745 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253274 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-trusted-ca\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253418 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-config\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253480 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hndw2\" (UniqueName: \"kubernetes.io/projected/014604aa-7ce2-48ed-b924-00ac8db7abb1-kube-api-access-hndw2\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253521 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253556 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-trusted-ca\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253586 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ngkf\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-kube-api-access-7ngkf\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253627 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xsws\" (UniqueName: \"kubernetes.io/projected/f0dd0c9a-f13a-4d9b-b812-da766c97e831-kube-api-access-4xsws\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253660 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43b9cf2c-e093-4cdf-b0d6-1a8279515176-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253738 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd8026a4-bc63-4912-94a8-897cf3c0bea6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253770 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-registry-certificates\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd60f8c1-b60d-4d72-b25d-1496429ebcae-metrics-tls\") pod \"dns-operator-744455d44c-qjwkt\" (UID: \"cd60f8c1-b60d-4d72-b25d-1496429ebcae\") " pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253835 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84w5l\" (UniqueName: \"kubernetes.io/projected/dcbd62f3-c933-4182-8172-3ef2fc3220ec-kube-api-access-84w5l\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.253971 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254022 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd8026a4-bc63-4912-94a8-897cf3c0bea6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254057 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254231 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254283 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-serving-cert\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254360 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/014604aa-7ce2-48ed-b924-00ac8db7abb1-images\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-registry-tls\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254525 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254556 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-service-ca\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254588 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ecdc8ba7-41d8-419a-ac31-52278e353468-machine-approver-tls\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254619 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254658 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43b9cf2c-e093-4cdf-b0d6-1a8279515176-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-bound-sa-token\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254759 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-service-ca-bundle\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254802 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dcbd62f3-c933-4182-8172-3ef2fc3220ec-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254831 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254859 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254887 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b7qc\" (UniqueName: \"kubernetes.io/projected/71fd30bb-8da6-4b86-8dd6-56274220a60b-kube-api-access-9b7qc\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254927 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzhgd\" (UniqueName: \"kubernetes.io/projected/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-kube-api-access-kzhgd\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.254972 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/014604aa-7ce2-48ed-b924-00ac8db7abb1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255007 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255054 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ctdw\" (UniqueName: \"kubernetes.io/projected/cd60f8c1-b60d-4d72-b25d-1496429ebcae-kube-api-access-7ctdw\") pod \"dns-operator-744455d44c-qjwkt\" (UID: \"cd60f8c1-b60d-4d72-b25d-1496429ebcae\") " pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255097 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255143 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255173 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/992baa6d-67aa-467a-8467-c97d64e69d44-installation-pull-secrets\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255210 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldmpl\" (UniqueName: \"kubernetes.io/projected/ecdc8ba7-41d8-419a-ac31-52278e353468-kube-api-access-ldmpl\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255238 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255260 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-serving-cert\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-dir\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255361 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zwj2\" (UniqueName: \"kubernetes.io/projected/fd8026a4-bc63-4912-94a8-897cf3c0bea6-kube-api-access-6zwj2\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255393 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0dd0c9a-f13a-4d9b-b812-da766c97e831-serving-cert\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255417 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-config\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255527 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255558 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ecdc8ba7-41d8-419a-ac31-52278e353468-auth-proxy-config\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255600 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-policies\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255662 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecdc8ba7-41d8-419a-ac31-52278e353468-config\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.255939 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:57.755920516 +0000 UTC m=+132.957297597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255942 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/014604aa-7ce2-48ed-b924-00ac8db7abb1-config\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.255989 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcbd62f3-c933-4182-8172-3ef2fc3220ec-serving-cert\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256014 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-client\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256040 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-config\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256071 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/992baa6d-67aa-467a-8467-c97d64e69d44-ca-trust-extracted\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256097 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5d9\" (UniqueName: \"kubernetes.io/projected/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-kube-api-access-ph5d9\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256123 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7h9f\" (UniqueName: \"kubernetes.io/projected/43b9cf2c-e093-4cdf-b0d6-1a8279515176-kube-api-access-f7h9f\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256143 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-config\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256165 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg6r4\" (UniqueName: \"kubernetes.io/projected/ceabfcec-18d7-4f41-af6c-b31edfcdbe80-kube-api-access-rg6r4\") pod \"downloads-7954f5f757-qkmml\" (UID: \"ceabfcec-18d7-4f41-af6c-b31edfcdbe80\") " pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.256206 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-ca\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.267120 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.280551 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.300252 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.320618 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.339852 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357489 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.357650 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:57.857625016 +0000 UTC m=+133.059001877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357754 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2f46\" (UniqueName: \"kubernetes.io/projected/c3615454-def8-4637-884b-a4c96fe468d6-kube-api-access-k2f46\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357784 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncsnp\" (UniqueName: \"kubernetes.io/projected/fbb99ef5-6a95-43f4-a404-1d4a14cf28e8-kube-api-access-ncsnp\") pod \"control-plane-machine-set-operator-78cbb6b69f-nx6lj\" (UID: \"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-policies\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecdc8ba7-41d8-419a-ac31-52278e353468-config\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357855 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6b9a502-ea7d-4dae-9539-89b710c40171-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357880 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcbd62f3-c933-4182-8172-3ef2fc3220ec-serving-cert\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357915 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-config\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357943 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-mountpoint-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.357978 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/992baa6d-67aa-467a-8467-c97d64e69d44-ca-trust-extracted\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358001 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f006d39c-146c-4461-aaa0-a81a6b7d535a-metrics-tls\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358030 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7h9f\" (UniqueName: \"kubernetes.io/projected/43b9cf2c-e093-4cdf-b0d6-1a8279515176-kube-api-access-f7h9f\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358053 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhpq4\" (UniqueName: \"kubernetes.io/projected/93b3d24c-56a8-485c-820d-7567ed1a6733-kube-api-access-fhpq4\") pod \"cluster-samples-operator-665b6dd947-kwfxl\" (UID: \"93b3d24c-56a8-485c-820d-7567ed1a6733\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358077 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctqp9\" (UniqueName: \"kubernetes.io/projected/c30fe471-6fb2-467a-b63e-2fc8f915c948-kube-api-access-ctqp9\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358098 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9css\" (UniqueName: \"kubernetes.io/projected/36d8e040-d731-4f82-ba5b-ace1145e15f6-kube-api-access-v9css\") pod \"multus-admission-controller-857f4d67dd-nnc2j\" (UID: \"36d8e040-d731-4f82-ba5b-ace1145e15f6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358131 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg6r4\" (UniqueName: \"kubernetes.io/projected/ceabfcec-18d7-4f41-af6c-b31edfcdbe80-kube-api-access-rg6r4\") pod \"downloads-7954f5f757-qkmml\" (UID: \"ceabfcec-18d7-4f41-af6c-b31edfcdbe80\") " pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358155 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-config\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358177 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmw9s\" (UniqueName: \"kubernetes.io/projected/82f6d6a9-f455-4dc5-bdfb-ac3616141bad-kube-api-access-dmw9s\") pod \"migrator-59844c95c7-xgmkj\" (UID: \"82f6d6a9-f455-4dc5-bdfb-ac3616141bad\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358202 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mflp6\" (UniqueName: \"kubernetes.io/projected/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-kube-api-access-mflp6\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358225 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40add3ff-6f8a-4b7d-a385-454665ca770c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358268 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w75w9\" (UniqueName: \"kubernetes.io/projected/abe3b323-2c4b-42a0-9357-a29866107e8a-kube-api-access-w75w9\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358294 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hndw2\" (UniqueName: \"kubernetes.io/projected/014604aa-7ce2-48ed-b924-00ac8db7abb1-kube-api-access-hndw2\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358317 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-service-ca\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358344 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40add3ff-6f8a-4b7d-a385-454665ca770c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358367 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-trusted-ca\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358394 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ngkf\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-kube-api-access-7ngkf\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358418 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-csi-data-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358459 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xsws\" (UniqueName: \"kubernetes.io/projected/f0dd0c9a-f13a-4d9b-b812-da766c97e831-kube-api-access-4xsws\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358483 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-config\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358505 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-config\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358532 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szhhg\" (UniqueName: \"kubernetes.io/projected/32009def-8f0b-4b71-961c-422ccb9aafea-kube-api-access-szhhg\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358558 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fbb99ef5-6a95-43f4-a404-1d4a14cf28e8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nx6lj\" (UID: \"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358568 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/992baa6d-67aa-467a-8467-c97d64e69d44-ca-trust-extracted\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358598 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/784b6e22-0844-494f-ac12-2fa16fccc21b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xsgd2\" (UID: \"784b6e22-0844-494f-ac12-2fa16fccc21b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358635 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358868 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecdc8ba7-41d8-419a-ac31-52278e353468-config\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.358941 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-registration-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359005 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vt4s\" (UniqueName: \"kubernetes.io/projected/a32f32fc-e1aa-402c-93ec-f73446955bd0-kube-api-access-2vt4s\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359022 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-config\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359034 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-policies\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359056 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-default-certificate\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359088 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359113 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359220 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/014604aa-7ce2-48ed-b924-00ac8db7abb1-images\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359304 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-config\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359348 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7psj4\" (UniqueName: \"kubernetes.io/projected/71308db0-b934-4c92-b90c-aaf1c6b1b41c-kube-api-access-7psj4\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359408 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-signing-cabundle\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359437 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-plugins-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359488 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-registry-tls\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359509 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6b9a502-ea7d-4dae-9539-89b710c40171-srv-cert\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359539 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg4z7\" (UniqueName: \"kubernetes.io/projected/a6da11ac-6aa4-43d3-927e-315d72618eb6-kube-api-access-qg4z7\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359585 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-config\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359658 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-service-ca\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359725 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ecdc8ba7-41d8-419a-ac31-52278e353468-machine-approver-tls\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359785 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359813 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-node-bootstrap-token\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.359947 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/abe3b323-2c4b-42a0-9357-a29866107e8a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360000 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f030a-c717-4277-b739-393ecb2081e4-secret-volume\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360030 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360087 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10048458-f2d3-4599-a7d2-727fe81b6006-auth-proxy-config\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360128 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dcbd62f3-c933-4182-8172-3ef2fc3220ec-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360190 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f006d39c-146c-4461-aaa0-a81a6b7d535a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360232 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360281 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360309 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b7qc\" (UniqueName: \"kubernetes.io/projected/71fd30bb-8da6-4b86-8dd6-56274220a60b-kube-api-access-9b7qc\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360332 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32009def-8f0b-4b71-961c-422ccb9aafea-webhook-cert\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360349 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/014604aa-7ce2-48ed-b924-00ac8db7abb1-images\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360365 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-service-ca\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/014604aa-7ce2-48ed-b924-00ac8db7abb1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzhgd\" (UniqueName: \"kubernetes.io/projected/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-kube-api-access-kzhgd\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360593 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74a07ef0-49b8-442a-b011-fc80b710cb80-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360615 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ctdw\" (UniqueName: \"kubernetes.io/projected/cd60f8c1-b60d-4d72-b25d-1496429ebcae-kube-api-access-7ctdw\") pod \"dns-operator-744455d44c-qjwkt\" (UID: \"cd60f8c1-b60d-4d72-b25d-1496429ebcae\") " pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/67bbd99c-59f5-4bb0-af30-d503cb341f2c-cert\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360633 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dcbd62f3-c933-4182-8172-3ef2fc3220ec-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360660 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-trusted-ca\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360787 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360819 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360849 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360875 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldmpl\" (UniqueName: \"kubernetes.io/projected/ecdc8ba7-41d8-419a-ac31-52278e353468-kube-api-access-ldmpl\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.360935 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-proxy-tls\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361020 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/abe3b323-2c4b-42a0-9357-a29866107e8a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361042 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-serving-cert\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361065 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361090 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0dd0c9a-f13a-4d9b-b812-da766c97e831-serving-cert\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361113 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-config\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361137 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-signing-key\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361162 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10048458-f2d3-4599-a7d2-727fe81b6006-images\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-metrics-certs\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361225 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361838 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.361960 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362327 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ecdc8ba7-41d8-419a-ac31-52278e353468-auth-proxy-config\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/014604aa-7ce2-48ed-b924-00ac8db7abb1-config\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362428 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-client\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362453 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxrft\" (UniqueName: \"kubernetes.io/projected/92c7bc44-a787-4fab-8213-e316dc8ca4dc-kube-api-access-nxrft\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362755 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362806 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt6nk\" (UniqueName: \"kubernetes.io/projected/40add3ff-6f8a-4b7d-a385-454665ca770c-kube-api-access-tt6nk\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362830 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/92c7bc44-a787-4fab-8213-e316dc8ca4dc-srv-cert\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362890 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-config\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362916 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5d9\" (UniqueName: \"kubernetes.io/projected/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-kube-api-access-ph5d9\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362940 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-serving-cert\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362959 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-ca\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.362959 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363060 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-schkb\" (UniqueName: \"kubernetes.io/projected/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-kube-api-access-schkb\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363103 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-trusted-ca\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74a07ef0-49b8-442a-b011-fc80b710cb80-config\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363161 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10048458-f2d3-4599-a7d2-727fe81b6006-proxy-tls\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dzf7\" (UniqueName: \"kubernetes.io/projected/10048458-f2d3-4599-a7d2-727fe81b6006-kube-api-access-2dzf7\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363195 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/014604aa-7ce2-48ed-b924-00ac8db7abb1-config\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363264 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp54v\" (UniqueName: \"kubernetes.io/projected/784b6e22-0844-494f-ac12-2fa16fccc21b-kube-api-access-hp54v\") pod \"package-server-manager-789f6589d5-xsgd2\" (UID: \"784b6e22-0844-494f-ac12-2fa16fccc21b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363294 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6da11ac-6aa4-43d3-927e-315d72618eb6-metrics-tls\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363388 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71308db0-b934-4c92-b90c-aaf1c6b1b41c-serving-cert\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363415 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-trusted-ca-bundle\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363441 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32009def-8f0b-4b71-961c-422ccb9aafea-apiservice-cert\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363468 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f8b5\" (UniqueName: \"kubernetes.io/projected/b6b9a502-ea7d-4dae-9539-89b710c40171-kube-api-access-6f8b5\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363499 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-stats-auth\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363522 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8c8\" (UniqueName: \"kubernetes.io/projected/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-kube-api-access-tp8c8\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363545 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-certs\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363642 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-config\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363719 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-config\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363732 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363876 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd8026a4-bc63-4912-94a8-897cf3c0bea6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363909 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43b9cf2c-e093-4cdf-b0d6-1a8279515176-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363967 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.363986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/93b3d24c-56a8-485c-820d-7567ed1a6733-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kwfxl\" (UID: \"93b3d24c-56a8-485c-820d-7567ed1a6733\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364011 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-registry-certificates\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ecdc8ba7-41d8-419a-ac31-52278e353468-machine-approver-tls\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364031 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd60f8c1-b60d-4d72-b25d-1496429ebcae-metrics-tls\") pod \"dns-operator-744455d44c-qjwkt\" (UID: \"cd60f8c1-b60d-4d72-b25d-1496429ebcae\") " pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364058 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-oauth-config\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364088 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84w5l\" (UniqueName: \"kubernetes.io/projected/dcbd62f3-c933-4182-8172-3ef2fc3220ec-kube-api-access-84w5l\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364195 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74a07ef0-49b8-442a-b011-fc80b710cb80-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364227 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd8026a4-bc63-4912-94a8-897cf3c0bea6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364242 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-trusted-ca\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364254 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-client-ca\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364284 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/abe3b323-2c4b-42a0-9357-a29866107e8a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364312 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-config\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364342 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-ca\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364389 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ecdc8ba7-41d8-419a-ac31-52278e353468-auth-proxy-config\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.364619 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:57.864603103 +0000 UTC m=+133.065979954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364774 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd8026a4-bc63-4912-94a8-897cf3c0bea6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.364988 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-registry-tls\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.365640 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-registry-certificates\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.365656 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.365757 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.365825 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-serving-cert\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.365871 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbkps\" (UniqueName: \"kubernetes.io/projected/886f030a-c717-4277-b739-393ecb2081e4-kube-api-access-kbkps\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.365903 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-socket-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.366072 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl2c8\" (UniqueName: \"kubernetes.io/projected/10a66e2c-cb51-4eec-864b-d8288ca3e281-kube-api-access-wl2c8\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.366264 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f030a-c717-4277-b739-393ecb2081e4-config-volume\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.366416 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7js6q\" (UniqueName: \"kubernetes.io/projected/58900711-350a-4753-888d-7a6ff520abd1-kube-api-access-7js6q\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.366511 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.366608 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.366987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58900711-350a-4753-888d-7a6ff520abd1-service-ca-bundle\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367039 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f006d39c-146c-4461-aaa0-a81a6b7d535a-trusted-ca\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367159 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-bound-sa-token\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43b9cf2c-e093-4cdf-b0d6-1a8279515176-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367406 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rfvl\" (UniqueName: \"kubernetes.io/projected/f006d39c-146c-4461-aaa0-a81a6b7d535a-kube-api-access-6rfvl\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367438 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/36d8e040-d731-4f82-ba5b-ace1145e15f6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nnc2j\" (UID: \"36d8e040-d731-4f82-ba5b-ace1145e15f6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367483 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-service-ca-bundle\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367557 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv9cj\" (UniqueName: \"kubernetes.io/projected/83593597-7147-4b61-be7c-e05b165bdebe-kube-api-access-zv9cj\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367750 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/32009def-8f0b-4b71-961c-422ccb9aafea-tmpfs\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367901 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c30fe471-6fb2-467a-b63e-2fc8f915c948-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367951 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-oauth-serving-cert\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.367970 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368046 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368119 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/992baa6d-67aa-467a-8467-c97d64e69d44-installation-pull-secrets\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368172 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-serving-cert\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368206 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368233 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6da11ac-6aa4-43d3-927e-315d72618eb6-config-volume\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368277 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/92c7bc44-a787-4fab-8213-e316dc8ca4dc-profile-collector-cert\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368301 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-service-ca-bundle\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368301 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43b9cf2c-e093-4cdf-b0d6-1a8279515176-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-dir\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368397 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zwj2\" (UniqueName: \"kubernetes.io/projected/fd8026a4-bc63-4912-94a8-897cf3c0bea6-kube-api-access-6zwj2\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368433 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slpp8\" (UniqueName: \"kubernetes.io/projected/67bbd99c-59f5-4bb0-af30-d503cb341f2c-kube-api-access-slpp8\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368465 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-dir\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.368876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.369818 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.369915 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.369995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd8026a4-bc63-4912-94a8-897cf3c0bea6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.369974 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcbd62f3-c933-4182-8172-3ef2fc3220ec-serving-cert\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.369952 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.370230 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0dd0c9a-f13a-4d9b-b812-da766c97e831-serving-cert\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.370380 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd60f8c1-b60d-4d72-b25d-1496429ebcae-metrics-tls\") pod \"dns-operator-744455d44c-qjwkt\" (UID: \"cd60f8c1-b60d-4d72-b25d-1496429ebcae\") " pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.370518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.371020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.371173 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/014604aa-7ce2-48ed-b924-00ac8db7abb1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.371191 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43b9cf2c-e093-4cdf-b0d6-1a8279515176-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.371631 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.372050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-serving-cert\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.372172 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/992baa6d-67aa-467a-8467-c97d64e69d44-installation-pull-secrets\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.373430 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0dd0c9a-f13a-4d9b-b812-da766c97e831-etcd-client\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.373536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-serving-cert\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.386608 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.400057 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.419807 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.440172 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.460042 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.470573 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.471003 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:57.970942964 +0000 UTC m=+133.172319825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471148 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10048458-f2d3-4599-a7d2-727fe81b6006-auth-proxy-config\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471190 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f006d39c-146c-4461-aaa0-a81a6b7d535a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471225 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32009def-8f0b-4b71-961c-422ccb9aafea-webhook-cert\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471256 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74a07ef0-49b8-442a-b011-fc80b710cb80-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471313 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/67bbd99c-59f5-4bb0-af30-d503cb341f2c-cert\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471361 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471405 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-proxy-tls\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471445 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/abe3b323-2c4b-42a0-9357-a29866107e8a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471495 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-signing-key\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471535 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-serving-cert\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471593 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10048458-f2d3-4599-a7d2-727fe81b6006-images\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-metrics-certs\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471652 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxrft\" (UniqueName: \"kubernetes.io/projected/92c7bc44-a787-4fab-8213-e316dc8ca4dc-kube-api-access-nxrft\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471697 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt6nk\" (UniqueName: \"kubernetes.io/projected/40add3ff-6f8a-4b7d-a385-454665ca770c-kube-api-access-tt6nk\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471723 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/92c7bc44-a787-4fab-8213-e316dc8ca4dc-srv-cert\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471766 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-serving-cert\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471786 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-schkb\" (UniqueName: \"kubernetes.io/projected/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-kube-api-access-schkb\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471825 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74a07ef0-49b8-442a-b011-fc80b710cb80-config\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10048458-f2d3-4599-a7d2-727fe81b6006-proxy-tls\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dzf7\" (UniqueName: \"kubernetes.io/projected/10048458-f2d3-4599-a7d2-727fe81b6006-kube-api-access-2dzf7\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471922 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp54v\" (UniqueName: \"kubernetes.io/projected/784b6e22-0844-494f-ac12-2fa16fccc21b-kube-api-access-hp54v\") pod \"package-server-manager-789f6589d5-xsgd2\" (UID: \"784b6e22-0844-494f-ac12-2fa16fccc21b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471945 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6da11ac-6aa4-43d3-927e-315d72618eb6-metrics-tls\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471970 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71308db0-b934-4c92-b90c-aaf1c6b1b41c-serving-cert\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.471990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-trusted-ca-bundle\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472013 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32009def-8f0b-4b71-961c-422ccb9aafea-apiservice-cert\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472039 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f8b5\" (UniqueName: \"kubernetes.io/projected/b6b9a502-ea7d-4dae-9539-89b710c40171-kube-api-access-6f8b5\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472064 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-stats-auth\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472084 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8c8\" (UniqueName: \"kubernetes.io/projected/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-kube-api-access-tp8c8\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472118 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/93b3d24c-56a8-485c-820d-7567ed1a6733-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kwfxl\" (UID: \"93b3d24c-56a8-485c-820d-7567ed1a6733\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472159 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-certs\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472180 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-oauth-config\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472211 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74a07ef0-49b8-442a-b011-fc80b710cb80-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472233 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-client-ca\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472254 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/abe3b323-2c4b-42a0-9357-a29866107e8a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472273 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-config\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbkps\" (UniqueName: \"kubernetes.io/projected/886f030a-c717-4277-b739-393ecb2081e4-kube-api-access-kbkps\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472330 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-socket-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472360 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl2c8\" (UniqueName: \"kubernetes.io/projected/10a66e2c-cb51-4eec-864b-d8288ca3e281-kube-api-access-wl2c8\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472390 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f030a-c717-4277-b739-393ecb2081e4-config-volume\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472422 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7js6q\" (UniqueName: \"kubernetes.io/projected/58900711-350a-4753-888d-7a6ff520abd1-kube-api-access-7js6q\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472475 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58900711-350a-4753-888d-7a6ff520abd1-service-ca-bundle\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472493 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f006d39c-146c-4461-aaa0-a81a6b7d535a-trusted-ca\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472533 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rfvl\" (UniqueName: \"kubernetes.io/projected/f006d39c-146c-4461-aaa0-a81a6b7d535a-kube-api-access-6rfvl\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472556 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/36d8e040-d731-4f82-ba5b-ace1145e15f6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nnc2j\" (UID: \"36d8e040-d731-4f82-ba5b-ace1145e15f6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472579 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv9cj\" (UniqueName: \"kubernetes.io/projected/83593597-7147-4b61-be7c-e05b165bdebe-kube-api-access-zv9cj\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472597 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/32009def-8f0b-4b71-961c-422ccb9aafea-tmpfs\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472618 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c30fe471-6fb2-467a-b63e-2fc8f915c948-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472642 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-oauth-serving-cert\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472664 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472709 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472740 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slpp8\" (UniqueName: \"kubernetes.io/projected/67bbd99c-59f5-4bb0-af30-d503cb341f2c-kube-api-access-slpp8\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472778 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6da11ac-6aa4-43d3-927e-315d72618eb6-config-volume\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472809 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/92c7bc44-a787-4fab-8213-e316dc8ca4dc-profile-collector-cert\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472840 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2f46\" (UniqueName: \"kubernetes.io/projected/c3615454-def8-4637-884b-a4c96fe468d6-kube-api-access-k2f46\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncsnp\" (UniqueName: \"kubernetes.io/projected/fbb99ef5-6a95-43f4-a404-1d4a14cf28e8-kube-api-access-ncsnp\") pod \"control-plane-machine-set-operator-78cbb6b69f-nx6lj\" (UID: \"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472886 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6b9a502-ea7d-4dae-9539-89b710c40171-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472904 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-mountpoint-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472961 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f006d39c-146c-4461-aaa0-a81a6b7d535a-metrics-tls\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.472996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhpq4\" (UniqueName: \"kubernetes.io/projected/93b3d24c-56a8-485c-820d-7567ed1a6733-kube-api-access-fhpq4\") pod \"cluster-samples-operator-665b6dd947-kwfxl\" (UID: \"93b3d24c-56a8-485c-820d-7567ed1a6733\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctqp9\" (UniqueName: \"kubernetes.io/projected/c30fe471-6fb2-467a-b63e-2fc8f915c948-kube-api-access-ctqp9\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473258 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9css\" (UniqueName: \"kubernetes.io/projected/36d8e040-d731-4f82-ba5b-ace1145e15f6-kube-api-access-v9css\") pod \"multus-admission-controller-857f4d67dd-nnc2j\" (UID: \"36d8e040-d731-4f82-ba5b-ace1145e15f6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmw9s\" (UniqueName: \"kubernetes.io/projected/82f6d6a9-f455-4dc5-bdfb-ac3616141bad-kube-api-access-dmw9s\") pod \"migrator-59844c95c7-xgmkj\" (UID: \"82f6d6a9-f455-4dc5-bdfb-ac3616141bad\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473634 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mflp6\" (UniqueName: \"kubernetes.io/projected/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-kube-api-access-mflp6\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473661 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40add3ff-6f8a-4b7d-a385-454665ca770c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473698 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w75w9\" (UniqueName: \"kubernetes.io/projected/abe3b323-2c4b-42a0-9357-a29866107e8a-kube-api-access-w75w9\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473740 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-service-ca\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.473759 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40add3ff-6f8a-4b7d-a385-454665ca770c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474052 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/32009def-8f0b-4b71-961c-422ccb9aafea-tmpfs\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474419 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-socket-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10048458-f2d3-4599-a7d2-727fe81b6006-auth-proxy-config\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40add3ff-6f8a-4b7d-a385-454665ca770c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474556 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-mountpoint-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474758 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-client-ca\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474821 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-csi-data-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474864 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-config\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474892 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-config\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474774 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-oauth-serving-cert\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474940 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/784b6e22-0844-494f-ac12-2fa16fccc21b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xsgd2\" (UID: \"784b6e22-0844-494f-ac12-2fa16fccc21b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.474980 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szhhg\" (UniqueName: \"kubernetes.io/projected/32009def-8f0b-4b71-961c-422ccb9aafea-kube-api-access-szhhg\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475045 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fbb99ef5-6a95-43f4-a404-1d4a14cf28e8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nx6lj\" (UID: \"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475091 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-registration-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vt4s\" (UniqueName: \"kubernetes.io/projected/a32f32fc-e1aa-402c-93ec-f73446955bd0-kube-api-access-2vt4s\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475158 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-default-certificate\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475183 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475208 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-config\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475233 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7psj4\" (UniqueName: \"kubernetes.io/projected/71308db0-b934-4c92-b90c-aaf1c6b1b41c-kube-api-access-7psj4\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475255 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475276 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-signing-cabundle\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475334 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-plugins-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475346 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74a07ef0-49b8-442a-b011-fc80b710cb80-config\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475383 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6b9a502-ea7d-4dae-9539-89b710c40171-srv-cert\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475412 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg4z7\" (UniqueName: \"kubernetes.io/projected/a6da11ac-6aa4-43d3-927e-315d72618eb6-kube-api-access-qg4z7\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475414 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-csi-data-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475549 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-node-bootstrap-token\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475591 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/abe3b323-2c4b-42a0-9357-a29866107e8a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475627 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f030a-c717-4277-b739-393ecb2081e4-secret-volume\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475659 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.475902 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-plugins-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.476104 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:57.976077891 +0000 UTC m=+133.177454782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.476147 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f006d39c-146c-4461-aaa0-a81a6b7d535a-trusted-ca\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.476188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-config\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.477155 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-oauth-config\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.477843 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40add3ff-6f8a-4b7d-a385-454665ca770c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.478250 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c3615454-def8-4637-884b-a4c96fe468d6-registration-dir\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.478252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f006d39c-146c-4461-aaa0-a81a6b7d535a-metrics-tls\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.478351 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-trusted-ca-bundle\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.478741 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-service-ca\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.479520 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.479916 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.480056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-config\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.480135 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-serving-cert\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.480763 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/abe3b323-2c4b-42a0-9357-a29866107e8a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.481732 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74a07ef0-49b8-442a-b011-fc80b710cb80-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.481995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/abe3b323-2c4b-42a0-9357-a29866107e8a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.482645 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71308db0-b934-4c92-b90c-aaf1c6b1b41c-serving-cert\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.482776 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fbb99ef5-6a95-43f4-a404-1d4a14cf28e8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nx6lj\" (UID: \"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.483359 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-proxy-tls\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.487641 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-config\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.500100 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.519565 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.539730 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.546920 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.559860 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.566888 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c30fe471-6fb2-467a-b63e-2fc8f915c948-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.576289 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.576592 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.076495315 +0000 UTC m=+133.277872186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.577344 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.577794 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.077776219 +0000 UTC m=+133.279153270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.579400 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.600380 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.609774 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6b9a502-ea7d-4dae-9539-89b710c40171-srv-cert\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.619724 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.626972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6b9a502-ea7d-4dae-9539-89b710c40171-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.628000 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/92c7bc44-a787-4fab-8213-e316dc8ca4dc-profile-collector-cert\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.631604 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f030a-c717-4277-b739-393ecb2081e4-secret-volume\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.639781 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.659937 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.679969 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.680125 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.180092092 +0000 UTC m=+133.381468993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.680274 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.680665 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.680804 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.180783802 +0000 UTC m=+133.382160703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.699886 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.720024 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.726641 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.740421 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.749841 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.760105 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.780612 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.781850 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.782036 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.282002936 +0000 UTC m=+133.483379797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.782447 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.783165 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.283143356 +0000 UTC m=+133.484520258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.786099 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/93b3d24c-56a8-485c-820d-7567ed1a6733-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kwfxl\" (UID: \"93b3d24c-56a8-485c-820d-7567ed1a6733\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.799772 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.819877 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.838478 4758 request.go:700] Waited for 1.0140131s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-admission-controller-secret&limit=500&resourceVersion=0 Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.840560 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.848100 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/36d8e040-d731-4f82-ba5b-ace1145e15f6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nnc2j\" (UID: \"36d8e040-d731-4f82-ba5b-ace1145e15f6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.860959 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.878969 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.883490 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.883727 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.383667211 +0000 UTC m=+133.585044072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.884133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.884550 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.384541875 +0000 UTC m=+133.585918726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.891988 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-stats-auth\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.900022 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.920557 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.939275 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.952303 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-default-certificate\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.959801 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.969161 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58900711-350a-4753-888d-7a6ff520abd1-metrics-certs\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.981419 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.985178 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.985915 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58900711-350a-4753-888d-7a6ff520abd1-service-ca-bundle\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:57 crc kubenswrapper[4758]: E1203 16:57:57.986052 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.486012747 +0000 UTC m=+133.687389608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:57 crc kubenswrapper[4758]: I1203 16:57:57.999799 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.020798 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.028821 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10048458-f2d3-4599-a7d2-727fe81b6006-images\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.039294 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.059231 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.068855 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10048458-f2d3-4599-a7d2-727fe81b6006-proxy-tls\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.079939 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.088353 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.089007 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.588980468 +0000 UTC m=+133.790357329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.090142 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/784b6e22-0844-494f-ac12-2fa16fccc21b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xsgd2\" (UID: \"784b6e22-0844-494f-ac12-2fa16fccc21b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.103753 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.120291 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.126458 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f030a-c717-4277-b739-393ecb2081e4-config-volume\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.140518 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.160791 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.167339 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-signing-cabundle\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.180860 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.189756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.189948 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.689909464 +0000 UTC m=+133.891286365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.190078 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.190659 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.690643084 +0000 UTC m=+133.892019975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.200307 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.220919 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.234024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-signing-key\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.238880 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.247102 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32009def-8f0b-4b71-961c-422ccb9aafea-webhook-cert\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.251286 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32009def-8f0b-4b71-961c-422ccb9aafea-apiservice-cert\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.259929 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.279927 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.291449 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.291791 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.791751745 +0000 UTC m=+133.993128646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.292408 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.293009 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.792983548 +0000 UTC m=+133.994360409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.300089 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.304979 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-config\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.320416 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.339306 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.349506 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-serving-cert\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.361266 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.380490 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.392787 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.393307 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.393434 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.893417202 +0000 UTC m=+134.094794063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.394004 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.394334 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.894326716 +0000 UTC m=+134.095703577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.408098 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.410133 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.420253 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.440199 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.460857 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.467883 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/92c7bc44-a787-4fab-8213-e316dc8ca4dc-srv-cert\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.473699 4758 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.473796 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-certs podName:10a66e2c-cb51-4eec-864b-d8288ca3e281 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.973772597 +0000 UTC m=+134.175149458 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-certs") pod "machine-config-server-hvg6p" (UID: "10a66e2c-cb51-4eec-864b-d8288ca3e281") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.473707 4758 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.473851 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a6da11ac-6aa4-43d3-927e-315d72618eb6-metrics-tls podName:a6da11ac-6aa4-43d3-927e-315d72618eb6 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.973839208 +0000 UTC m=+134.175216299 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a6da11ac-6aa4-43d3-927e-315d72618eb6-metrics-tls") pod "dns-default-c9b57" (UID: "a6da11ac-6aa4-43d3-927e-315d72618eb6") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.476919 4758 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.477006 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-node-bootstrap-token podName:10a66e2c-cb51-4eec-864b-d8288ca3e281 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.976986952 +0000 UTC m=+134.178363823 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-node-bootstrap-token") pod "machine-config-server-hvg6p" (UID: "10a66e2c-cb51-4eec-864b-d8288ca3e281") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.478157 4758 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.478276 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a6da11ac-6aa4-43d3-927e-315d72618eb6-config-volume podName:a6da11ac-6aa4-43d3-927e-315d72618eb6 nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.978248776 +0000 UTC m=+134.179625637 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a6da11ac-6aa4-43d3-927e-315d72618eb6-config-volume") pod "dns-default-c9b57" (UID: "a6da11ac-6aa4-43d3-927e-315d72618eb6") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.478777 4758 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.478823 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67bbd99c-59f5-4bb0-af30-d503cb341f2c-cert podName:67bbd99c-59f5-4bb0-af30-d503cb341f2c nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.978811552 +0000 UTC m=+134.180188413 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/67bbd99c-59f5-4bb0-af30-d503cb341f2c-cert") pod "ingress-canary-dcg47" (UID: "67bbd99c-59f5-4bb0-af30-d503cb341f2c") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.479991 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.495567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.495770 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.995750146 +0000 UTC m=+134.197127007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.496113 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.496609 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:58.996599018 +0000 UTC m=+134.197975879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.499613 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.520369 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.539590 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.560084 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.579633 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.597768 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.598074 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.098036959 +0000 UTC m=+134.299413840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.599055 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.599135 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.599773 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.099754485 +0000 UTC m=+134.301131346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.620144 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.641261 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.659576 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.680035 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.700435 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.700706 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.701021 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.200969179 +0000 UTC m=+134.402346090 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.701639 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.702177 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.2021378 +0000 UTC m=+134.403514671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.740400 4758 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.743666 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8rrt\" (UniqueName: \"kubernetes.io/projected/e9ea76f4-50d8-416a-85dd-897f08112fa3-kube-api-access-k8rrt\") pod \"apiserver-7bbb656c7d-7s4h7\" (UID: \"e9ea76f4-50d8-416a-85dd-897f08112fa3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.779391 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-978qs\" (UniqueName: \"kubernetes.io/projected/f4a6fd6b-4562-42ba-85d6-710a8f6bb98b-kube-api-access-978qs\") pod \"apiserver-76f77b778f-f4wqg\" (UID: \"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b\") " pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.780447 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.783936 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.801233 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.803408 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.803541 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.303514408 +0000 UTC m=+134.504891269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.804185 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.804874 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.304839754 +0000 UTC m=+134.506216655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.820927 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.838769 4758 request.go:700] Waited for 1.724442524s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&limit=500&resourceVersion=0 Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.841944 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.861208 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.876242 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.881535 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.905874 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:58 crc kubenswrapper[4758]: E1203 16:57:58.909982 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.409950843 +0000 UTC m=+134.611327724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.936713 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg6r4\" (UniqueName: \"kubernetes.io/projected/ceabfcec-18d7-4f41-af6c-b31edfcdbe80-kube-api-access-rg6r4\") pod \"downloads-7954f5f757-qkmml\" (UID: \"ceabfcec-18d7-4f41-af6c-b31edfcdbe80\") " pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.951619 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.955326 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7h9f\" (UniqueName: \"kubernetes.io/projected/43b9cf2c-e093-4cdf-b0d6-1a8279515176-kube-api-access-f7h9f\") pod \"kube-storage-version-migrator-operator-b67b599dd-gsnzz\" (UID: \"43b9cf2c-e093-4cdf-b0d6-1a8279515176\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:58 crc kubenswrapper[4758]: I1203 16:57:58.976559 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xsws\" (UniqueName: \"kubernetes.io/projected/f0dd0c9a-f13a-4d9b-b812-da766c97e831-kube-api-access-4xsws\") pod \"etcd-operator-b45778765-hqwcv\" (UID: \"f0dd0c9a-f13a-4d9b-b812-da766c97e831\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.005902 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hndw2\" (UniqueName: \"kubernetes.io/projected/014604aa-7ce2-48ed-b924-00ac8db7abb1-kube-api-access-hndw2\") pod \"machine-api-operator-5694c8668f-7lnzm\" (UID: \"014604aa-7ce2-48ed-b924-00ac8db7abb1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.008278 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-node-bootstrap-token\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.008369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/67bbd99c-59f5-4bb0-af30-d503cb341f2c-cert\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.008414 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.008509 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6da11ac-6aa4-43d3-927e-315d72618eb6-metrics-tls\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.008560 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-certs\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.008724 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6da11ac-6aa4-43d3-927e-315d72618eb6-config-volume\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.009611 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6da11ac-6aa4-43d3-927e-315d72618eb6-config-volume\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.011507 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/67bbd99c-59f5-4bb0-af30-d503cb341f2c-cert\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.011902 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.511884996 +0000 UTC m=+134.713261857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.013735 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-node-bootstrap-token\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.021314 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6da11ac-6aa4-43d3-927e-315d72618eb6-metrics-tls\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.021741 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10a66e2c-cb51-4eec-864b-d8288ca3e281-certs\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.024098 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f4wqg"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.031819 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ngkf\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-kube-api-access-7ngkf\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.041266 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b7qc\" (UniqueName: \"kubernetes.io/projected/71fd30bb-8da6-4b86-8dd6-56274220a60b-kube-api-access-9b7qc\") pod \"oauth-openshift-558db77b4-nqrwv\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.057320 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldmpl\" (UniqueName: \"kubernetes.io/projected/ecdc8ba7-41d8-419a-ac31-52278e353468-kube-api-access-ldmpl\") pod \"machine-approver-56656f9798-gz4q6\" (UID: \"ecdc8ba7-41d8-419a-ac31-52278e353468\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.080652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzhgd\" (UniqueName: \"kubernetes.io/projected/c138db90-10f2-4e2b-8cb3-8b368cbc1a7c-kube-api-access-kzhgd\") pod \"console-operator-58897d9998-4l6l4\" (UID: \"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c\") " pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.081011 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.089283 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.101104 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.101469 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ctdw\" (UniqueName: \"kubernetes.io/projected/cd60f8c1-b60d-4d72-b25d-1496429ebcae-kube-api-access-7ctdw\") pod \"dns-operator-744455d44c-qjwkt\" (UID: \"cd60f8c1-b60d-4d72-b25d-1496429ebcae\") " pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:59 crc kubenswrapper[4758]: W1203 16:57:59.106243 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9ea76f4_50d8_416a_85dd_897f08112fa3.slice/crio-b36b6b4150813c7b8491629356cd455c4f989205d73240ec8883f3f1ddfe5cde WatchSource:0}: Error finding container b36b6b4150813c7b8491629356cd455c4f989205d73240ec8883f3f1ddfe5cde: Status 404 returned error can't find the container with id b36b6b4150813c7b8491629356cd455c4f989205d73240ec8883f3f1ddfe5cde Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.109737 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.109896 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.609867004 +0000 UTC m=+134.811243865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.110366 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.110993 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.610979594 +0000 UTC m=+134.812356655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.116659 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5d9\" (UniqueName: \"kubernetes.io/projected/a35fa7d5-51a2-4e11-b805-e79d0ddbafa2-kube-api-access-ph5d9\") pod \"authentication-operator-69f744f599-8dmlt\" (UID: \"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.137316 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84w5l\" (UniqueName: \"kubernetes.io/projected/dcbd62f3-c933-4182-8172-3ef2fc3220ec-kube-api-access-84w5l\") pod \"openshift-config-operator-7777fb866f-2xvlb\" (UID: \"dcbd62f3-c933-4182-8172-3ef2fc3220ec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.152634 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qkmml"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.160949 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/813a5ac2-3ab0-4b77-9017-f79f849e7ce1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jmjcs\" (UID: \"813a5ac2-3ab0-4b77-9017-f79f849e7ce1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:59 crc kubenswrapper[4758]: W1203 16:57:59.169335 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podceabfcec_18d7_4f41_af6c_b31edfcdbe80.slice/crio-2bb3948e7dd4b5dabcd31128d94181f2d93bb57fe1983ddc15ff771a39c50176 WatchSource:0}: Error finding container 2bb3948e7dd4b5dabcd31128d94181f2d93bb57fe1983ddc15ff771a39c50176: Status 404 returned error can't find the container with id 2bb3948e7dd4b5dabcd31128d94181f2d93bb57fe1983ddc15ff771a39c50176 Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.175812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-bound-sa-token\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.201903 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zwj2\" (UniqueName: \"kubernetes.io/projected/fd8026a4-bc63-4912-94a8-897cf3c0bea6-kube-api-access-6zwj2\") pod \"openshift-apiserver-operator-796bbdcf4f-smpbk\" (UID: \"fd8026a4-bc63-4912-94a8-897cf3c0bea6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.211989 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.212238 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.212768 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.712748683 +0000 UTC m=+134.914125544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.216230 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhpq4\" (UniqueName: \"kubernetes.io/projected/93b3d24c-56a8-485c-820d-7567ed1a6733-kube-api-access-fhpq4\") pod \"cluster-samples-operator-665b6dd947-kwfxl\" (UID: \"93b3d24c-56a8-485c-820d-7567ed1a6733\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.220832 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.229239 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.235765 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctqp9\" (UniqueName: \"kubernetes.io/projected/c30fe471-6fb2-467a-b63e-2fc8f915c948-kube-api-access-ctqp9\") pod \"route-controller-manager-6576b87f9c-4nzm7\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.239021 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.255370 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slpp8\" (UniqueName: \"kubernetes.io/projected/67bbd99c-59f5-4bb0-af30-d503cb341f2c-kube-api-access-slpp8\") pod \"ingress-canary-dcg47\" (UID: \"67bbd99c-59f5-4bb0-af30-d503cb341f2c\") " pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.265909 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.269603 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.276758 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.277729 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncsnp\" (UniqueName: \"kubernetes.io/projected/fbb99ef5-6a95-43f4-a404-1d4a14cf28e8-kube-api-access-ncsnp\") pod \"control-plane-machine-set-operator-78cbb6b69f-nx6lj\" (UID: \"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.283715 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.289987 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dcg47" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.293091 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.296291 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv9cj\" (UniqueName: \"kubernetes.io/projected/83593597-7147-4b61-be7c-e05b165bdebe-kube-api-access-zv9cj\") pod \"marketplace-operator-79b997595-r8sqq\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.314046 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.314464 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.81444973 +0000 UTC m=+135.015826591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.319419 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9css\" (UniqueName: \"kubernetes.io/projected/36d8e040-d731-4f82-ba5b-ace1145e15f6-kube-api-access-v9css\") pod \"multus-admission-controller-857f4d67dd-nnc2j\" (UID: \"36d8e040-d731-4f82-ba5b-ace1145e15f6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.319474 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7lnzm"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.336374 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f006d39c-146c-4461-aaa0-a81a6b7d535a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.359093 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74a07ef0-49b8-442a-b011-fc80b710cb80-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wm6jh\" (UID: \"74a07ef0-49b8-442a-b011-fc80b710cb80\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.399489 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-schkb\" (UniqueName: \"kubernetes.io/projected/7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63-kube-api-access-schkb\") pod \"machine-config-controller-84d6567774-x976j\" (UID: \"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.413080 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.415635 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.416250 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:57:59.91623042 +0000 UTC m=+135.117607281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.430460 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbkps\" (UniqueName: \"kubernetes.io/projected/886f030a-c717-4277-b739-393ecb2081e4-kube-api-access-kbkps\") pod \"collect-profiles-29413005-7jdnm\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.433352 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.438065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.455401 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.459638 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl2c8\" (UniqueName: \"kubernetes.io/projected/10a66e2c-cb51-4eec-864b-d8288ca3e281-kube-api-access-wl2c8\") pod \"machine-config-server-hvg6p\" (UID: \"10a66e2c-cb51-4eec-864b-d8288ca3e281\") " pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.475253 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.475418 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmw9s\" (UniqueName: \"kubernetes.io/projected/82f6d6a9-f455-4dc5-bdfb-ac3616141bad-kube-api-access-dmw9s\") pod \"migrator-59844c95c7-xgmkj\" (UID: \"82f6d6a9-f455-4dc5-bdfb-ac3616141bad\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.482216 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.488672 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nqrwv"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.512436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7js6q\" (UniqueName: \"kubernetes.io/projected/58900711-350a-4753-888d-7a6ff520abd1-kube-api-access-7js6q\") pod \"router-default-5444994796-jxqhr\" (UID: \"58900711-350a-4753-888d-7a6ff520abd1\") " pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.519027 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mflp6\" (UniqueName: \"kubernetes.io/projected/bd7d3310-f1bb-4557-a852-bd4f1f72eb60-kube-api-access-mflp6\") pod \"service-ca-9c57cc56f-5hdt9\" (UID: \"bd7d3310-f1bb-4557-a852-bd4f1f72eb60\") " pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.519073 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.519533 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.01951598 +0000 UTC m=+135.220892841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.521959 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.529013 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.539380 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dzf7\" (UniqueName: \"kubernetes.io/projected/10048458-f2d3-4599-a7d2-727fe81b6006-kube-api-access-2dzf7\") pod \"machine-config-operator-74547568cd-srplx\" (UID: \"10048458-f2d3-4599-a7d2-727fe81b6006\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.556014 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp54v\" (UniqueName: \"kubernetes.io/projected/784b6e22-0844-494f-ac12-2fa16fccc21b-kube-api-access-hp54v\") pod \"package-server-manager-789f6589d5-xsgd2\" (UID: \"784b6e22-0844-494f-ac12-2fa16fccc21b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.556617 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.557501 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hqwcv"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.577291 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg4z7\" (UniqueName: \"kubernetes.io/projected/a6da11ac-6aa4-43d3-927e-315d72618eb6-kube-api-access-qg4z7\") pod \"dns-default-c9b57\" (UID: \"a6da11ac-6aa4-43d3-927e-315d72618eb6\") " pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.577708 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hvg6p" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.577767 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.597798 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljss6\" (UID: \"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.619585 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rfvl\" (UniqueName: \"kubernetes.io/projected/f006d39c-146c-4461-aaa0-a81a6b7d535a-kube-api-access-6rfvl\") pod \"ingress-operator-5b745b69d9-6r465\" (UID: \"f006d39c-146c-4461-aaa0-a81a6b7d535a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.619962 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.620314 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.120280062 +0000 UTC m=+135.321656923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.620644 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.621150 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.121126304 +0000 UTC m=+135.322503165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.623117 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qjwkt"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.654671 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szhhg\" (UniqueName: \"kubernetes.io/projected/32009def-8f0b-4b71-961c-422ccb9aafea-kube-api-access-szhhg\") pod \"packageserver-d55dfcdfc-t4xg7\" (UID: \"32009def-8f0b-4b71-961c-422ccb9aafea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.656437 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.673855 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f8b5\" (UniqueName: \"kubernetes.io/projected/b6b9a502-ea7d-4dae-9539-89b710c40171-kube-api-access-6f8b5\") pod \"olm-operator-6b444d44fb-lxkjl\" (UID: \"b6b9a502-ea7d-4dae-9539-89b710c40171\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.694613 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8c8\" (UniqueName: \"kubernetes.io/projected/bf1a9d85-21f7-446e-9ca3-e2d3d9010130-kube-api-access-tp8c8\") pod \"service-ca-operator-777779d784-xlvqm\" (UID: \"bf1a9d85-21f7-446e-9ca3-e2d3d9010130\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.714959 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt6nk\" (UniqueName: \"kubernetes.io/projected/40add3ff-6f8a-4b7d-a385-454665ca770c-kube-api-access-tt6nk\") pod \"openshift-controller-manager-operator-756b6f6bc6-hv699\" (UID: \"40add3ff-6f8a-4b7d-a385-454665ca770c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.721252 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.721657 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.721792 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.221763423 +0000 UTC m=+135.423140314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.722015 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.722537 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.222521864 +0000 UTC m=+135.423898755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.746480 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.753661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vt4s\" (UniqueName: \"kubernetes.io/projected/a32f32fc-e1aa-402c-93ec-f73446955bd0-kube-api-access-2vt4s\") pod \"console-f9d7485db-kzv6z\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.760747 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.767373 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.778981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7psj4\" (UniqueName: \"kubernetes.io/projected/71308db0-b934-4c92-b90c-aaf1c6b1b41c-kube-api-access-7psj4\") pod \"controller-manager-879f6c89f-4f6xq\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.779346 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxrft\" (UniqueName: \"kubernetes.io/projected/92c7bc44-a787-4fab-8213-e316dc8ca4dc-kube-api-access-nxrft\") pod \"catalog-operator-68c6474976-xj7gk\" (UID: \"92c7bc44-a787-4fab-8213-e316dc8ca4dc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.779337 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/abe3b323-2c4b-42a0-9357-a29866107e8a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.782958 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w75w9\" (UniqueName: \"kubernetes.io/projected/abe3b323-2c4b-42a0-9357-a29866107e8a-kube-api-access-w75w9\") pod \"cluster-image-registry-operator-dc59b4c8b-ntv4j\" (UID: \"abe3b323-2c4b-42a0-9357-a29866107e8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.786825 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2f46\" (UniqueName: \"kubernetes.io/projected/c3615454-def8-4637-884b-a4c96fe468d6-kube-api-access-k2f46\") pod \"csi-hostpathplugin-2gw5x\" (UID: \"c3615454-def8-4637-884b-a4c96fe468d6\") " pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:59 crc kubenswrapper[4758]: W1203 16:57:59.791459 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71fd30bb_8da6_4b86_8dd6_56274220a60b.slice/crio-6d479aa4ca446f444bd54948d86af20cab3ffb228cced346168c96d3796b0dbd WatchSource:0}: Error finding container 6d479aa4ca446f444bd54948d86af20cab3ffb228cced346168c96d3796b0dbd: Status 404 returned error can't find the container with id 6d479aa4ca446f444bd54948d86af20cab3ffb228cced346168c96d3796b0dbd Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.799017 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:57:59 crc kubenswrapper[4758]: W1203 16:57:59.801621 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0dd0c9a_f13a_4d9b_b812_da766c97e831.slice/crio-6b687f30e0baa0ff5ea54a25d7aa1d7f86d81f92b6f0e80d4aef8d853960d22d WatchSource:0}: Error finding container 6b687f30e0baa0ff5ea54a25d7aa1d7f86d81f92b6f0e80d4aef8d853960d22d: Status 404 returned error can't find the container with id 6b687f30e0baa0ff5ea54a25d7aa1d7f86d81f92b6f0e80d4aef8d853960d22d Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.820439 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.821080 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.828770 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.837391 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.337353593 +0000 UTC m=+135.538730454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.846003 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.847869 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.864082 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.872111 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c9b57" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.894094 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" event={"ID":"f0dd0c9a-f13a-4d9b-b812-da766c97e831","Type":"ContainerStarted","Data":"6b687f30e0baa0ff5ea54a25d7aa1d7f86d81f92b6f0e80d4aef8d853960d22d"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.896031 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dcg47"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.916412 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" event={"ID":"e9ea76f4-50d8-416a-85dd-897f08112fa3","Type":"ContainerStarted","Data":"b36b6b4150813c7b8491629356cd455c4f989205d73240ec8883f3f1ddfe5cde"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.917563 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.918890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" event={"ID":"ecdc8ba7-41d8-419a-ac31-52278e353468","Type":"ContainerStarted","Data":"832da6377107b80a84691e59441659aa039629fc3962fe2db49eeaef6a702754"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.921170 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" event={"ID":"71fd30bb-8da6-4b86-8dd6-56274220a60b","Type":"ContainerStarted","Data":"6d479aa4ca446f444bd54948d86af20cab3ffb228cced346168c96d3796b0dbd"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.922749 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qkmml" event={"ID":"ceabfcec-18d7-4f41-af6c-b31edfcdbe80","Type":"ContainerStarted","Data":"775800f23ea1ecbc69ffb09584d7fb54b7b745299571ed74ad55b340748366b2"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.922834 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qkmml" event={"ID":"ceabfcec-18d7-4f41-af6c-b31edfcdbe80","Type":"ContainerStarted","Data":"2bb3948e7dd4b5dabcd31128d94181f2d93bb57fe1983ddc15ff771a39c50176"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.923035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.923490 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8dmlt"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.923737 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" event={"ID":"fd8026a4-bc63-4912-94a8-897cf3c0bea6","Type":"ContainerStarted","Data":"741758d84b3439c9bf38fdaf9a2782c3dd4df525f96515a7b91ef5b2d8c2f1b6"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.924606 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4l6l4"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.925047 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-qkmml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.925088 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qkmml" podUID="ceabfcec-18d7-4f41-af6c-b31edfcdbe80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.925324 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" event={"ID":"cd60f8c1-b60d-4d72-b25d-1496429ebcae","Type":"ContainerStarted","Data":"41f785af489ef9949108528da4e0aaa6f047dc6360e30809c448e4dd72cfd7b5"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.930533 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:57:59 crc kubenswrapper[4758]: E1203 16:57:59.931162 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.431136338 +0000 UTC m=+135.632513199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.934915 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" event={"ID":"014604aa-7ce2-48ed-b924-00ac8db7abb1","Type":"ContainerStarted","Data":"3602c1a796d5567c61dd6b2f1fc4eb8d296b9799cf831632ceb1625f9024473d"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.937615 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" event={"ID":"43b9cf2c-e093-4cdf-b0d6-1a8279515176","Type":"ContainerStarted","Data":"144f18423fa7b3c4c9e79661e2361ce27c2ec0b216be4d6e4f10b27eb85bf1d2"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.939695 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs"] Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.939722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" event={"ID":"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b","Type":"ContainerStarted","Data":"63f5f7fbab39651f24d457433e61a878e980bcc17acaae6eea045918519993c8"} Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.994160 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:57:59 crc kubenswrapper[4758]: I1203 16:57:59.995821 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" Dec 03 16:57:59 crc kubenswrapper[4758]: W1203 16:57:59.998525 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67bbd99c_59f5_4bb0_af30_d503cb341f2c.slice/crio-2ea15396b98bedd95bd5220c2b53e101109df0e92a5b173defeb880473003a10 WatchSource:0}: Error finding container 2ea15396b98bedd95bd5220c2b53e101109df0e92a5b173defeb880473003a10: Status 404 returned error can't find the container with id 2ea15396b98bedd95bd5220c2b53e101109df0e92a5b173defeb880473003a10 Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.004650 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.027748 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.031875 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.032751 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.532730472 +0000 UTC m=+135.734107333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.134498 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.135735 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.635709514 +0000 UTC m=+135.837086375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.162754 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nnc2j"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.237005 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.237466 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6"] Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.237544 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.737522544 +0000 UTC m=+135.938899405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.340367 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.340833 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.840814834 +0000 UTC m=+136.042191695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.441472 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.441968 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:00.941944686 +0000 UTC m=+136.143321547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.546220 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.546569 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.547004 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.046985033 +0000 UTC m=+136.248361894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: W1203 16:58:00.560051 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36d8e040_d731_4f82_ba5b_ace1145e15f6.slice/crio-6354feb486abb97b7ec2ad5f0b1e750a35a90808077bf8502198bbfbd1f8d5ae WatchSource:0}: Error finding container 6354feb486abb97b7ec2ad5f0b1e750a35a90808077bf8502198bbfbd1f8d5ae: Status 404 returned error can't find the container with id 6354feb486abb97b7ec2ad5f0b1e750a35a90808077bf8502198bbfbd1f8d5ae Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.591276 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qkmml" podStartSLOduration=117.59125244 podStartE2EDuration="1m57.59125244s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:00.58306167 +0000 UTC m=+135.784438531" watchObservedRunningTime="2025-12-03 16:58:00.59125244 +0000 UTC m=+135.792629291" Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.596756 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.601182 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm"] Dec 03 16:58:00 crc kubenswrapper[4758]: W1203 16:58:00.617826 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacb4a12e_76ef_4e9a_b6e4_18e4fe4ff319.slice/crio-43944baefe97f828d73eee31ace175e1a8e4b72f2733a679e8eca202e878cf0b WatchSource:0}: Error finding container 43944baefe97f828d73eee31ace175e1a8e4b72f2733a679e8eca202e878cf0b: Status 404 returned error can't find the container with id 43944baefe97f828d73eee31ace175e1a8e4b72f2733a679e8eca202e878cf0b Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.652295 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.652707 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.152618375 +0000 UTC m=+136.353995236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.654748 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.655316 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.155294727 +0000 UTC m=+136.356671588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: W1203 16:58:00.700733 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82f6d6a9_f455_4dc5_bdfb_ac3616141bad.slice/crio-bfc9de1a16fe37196bedb8c686f7357cb6639417c96c4b7e611ec9c1ed62fd6e WatchSource:0}: Error finding container bfc9de1a16fe37196bedb8c686f7357cb6639417c96c4b7e611ec9c1ed62fd6e: Status 404 returned error can't find the container with id bfc9de1a16fe37196bedb8c686f7357cb6639417c96c4b7e611ec9c1ed62fd6e Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.756526 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.757018 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.256992235 +0000 UTC m=+136.458369086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: W1203 16:58:00.782624 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod886f030a_c717_4277_b739_393ecb2081e4.slice/crio-963b01f6bff28045993d91540cfc52ed8041ca3f8e4835d3d714afe976826976 WatchSource:0}: Error finding container 963b01f6bff28045993d91540cfc52ed8041ca3f8e4835d3d714afe976826976: Status 404 returned error can't find the container with id 963b01f6bff28045993d91540cfc52ed8041ca3f8e4835d3d714afe976826976 Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.801758 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.804557 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.825261 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5hdt9"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.838756 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.859154 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.859732 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.359710879 +0000 UTC m=+136.561087740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.875572 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.921423 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8sqq"] Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.960710 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:00 crc kubenswrapper[4758]: E1203 16:58:00.961277 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.461258643 +0000 UTC m=+136.662635504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:00 crc kubenswrapper[4758]: W1203 16:58:00.976949 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92c7bc44_a787_4fab_8213_e316dc8ca4dc.slice/crio-3f9c932b9520a4d41370521851a908c5201f4cbce948e65a8ba504c1a1b95c5a WatchSource:0}: Error finding container 3f9c932b9520a4d41370521851a908c5201f4cbce948e65a8ba504c1a1b95c5a: Status 404 returned error can't find the container with id 3f9c932b9520a4d41370521851a908c5201f4cbce948e65a8ba504c1a1b95c5a Dec 03 16:58:00 crc kubenswrapper[4758]: I1203 16:58:00.983510 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2gw5x"] Dec 03 16:58:00 crc kubenswrapper[4758]: W1203 16:58:00.991123 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83593597_7147_4b61_be7c_e05b165bdebe.slice/crio-128fc95a96bc235f23f52262de8749879b7e5255b470a20ef23180daf4dcfdd8 WatchSource:0}: Error finding container 128fc95a96bc235f23f52262de8749879b7e5255b470a20ef23180daf4dcfdd8: Status 404 returned error can't find the container with id 128fc95a96bc235f23f52262de8749879b7e5255b470a20ef23180daf4dcfdd8 Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.034378 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.048592 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" event={"ID":"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319","Type":"ContainerStarted","Data":"43944baefe97f828d73eee31ace175e1a8e4b72f2733a679e8eca202e878cf0b"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.063554 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.063957 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.563943276 +0000 UTC m=+136.765320137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.064329 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.065808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jxqhr" event={"ID":"58900711-350a-4753-888d-7a6ff520abd1","Type":"ContainerStarted","Data":"fbcb404f091a5711ab55a6b73a80a97029e140da92482d08c630b6c3c45bb778"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.068216 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c9b57"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.074105 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.074165 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" event={"ID":"ecdc8ba7-41d8-419a-ac31-52278e353468","Type":"ContainerStarted","Data":"331e4b5c6112a96f85ccf5686e5dba63be21bba5f857a74b350738543752af93"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.106953 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6r465"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.112728 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.112795 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x976j"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.166518 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.167018 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.666994499 +0000 UTC m=+136.868371370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.168342 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kzv6z"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.168375 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" event={"ID":"43b9cf2c-e093-4cdf-b0d6-1a8279515176","Type":"ContainerStarted","Data":"47c9d1a9f6fa530e4b611c44893bb2e963312fc0243950f1bccd6578c7328dde"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.168406 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.176309 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-srplx"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.190475 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4a6fd6b-4562-42ba-85d6-710a8f6bb98b" containerID="702e79ae1cb4270080f02d9b43535337f57c425536bf830adb61e7db17a000a2" exitCode=0 Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.190586 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" event={"ID":"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b","Type":"ContainerDied","Data":"702e79ae1cb4270080f02d9b43535337f57c425536bf830adb61e7db17a000a2"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.215522 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dcg47" event={"ID":"67bbd99c-59f5-4bb0-af30-d503cb341f2c","Type":"ContainerStarted","Data":"2ea15396b98bedd95bd5220c2b53e101109df0e92a5b173defeb880473003a10"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.225389 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.237012 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hvg6p" event={"ID":"10a66e2c-cb51-4eec-864b-d8288ca3e281","Type":"ContainerStarted","Data":"7316f694f1e578e12504ee8d95fe33684828ddcb32620e99844110ce5268274b"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.269425 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.271018 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.770997608 +0000 UTC m=+136.972374469 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.277391 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" event={"ID":"e9ea76f4-50d8-416a-85dd-897f08112fa3","Type":"ContainerDied","Data":"5dd4b899301bc78e89174276b524affc1449066649a722b0085a1af1f26d523f"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.278664 4758 generic.go:334] "Generic (PLEG): container finished" podID="e9ea76f4-50d8-416a-85dd-897f08112fa3" containerID="5dd4b899301bc78e89174276b524affc1449066649a722b0085a1af1f26d523f" exitCode=0 Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.281301 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" event={"ID":"82f6d6a9-f455-4dc5-bdfb-ac3616141bad","Type":"ContainerStarted","Data":"bfc9de1a16fe37196bedb8c686f7357cb6639417c96c4b7e611ec9c1ed62fd6e"} Dec 03 16:58:01 crc kubenswrapper[4758]: W1203 16:58:01.282822 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf006d39c_146c_4461_aaa0_a81a6b7d535a.slice/crio-c89070803b44fd7c2f596ec17503db6247e4332187be16e2ce39900ce6433434 WatchSource:0}: Error finding container c89070803b44fd7c2f596ec17503db6247e4332187be16e2ce39900ce6433434: Status 404 returned error can't find the container with id c89070803b44fd7c2f596ec17503db6247e4332187be16e2ce39900ce6433434 Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.283640 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.296023 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" event={"ID":"36d8e040-d731-4f82-ba5b-ace1145e15f6","Type":"ContainerStarted","Data":"6354feb486abb97b7ec2ad5f0b1e750a35a90808077bf8502198bbfbd1f8d5ae"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.300153 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" event={"ID":"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c","Type":"ContainerStarted","Data":"13b13ed46724f08ca2573bc1074541e2bfc58004f5d915afd470b1c1bccbaac1"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.309187 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" event={"ID":"886f030a-c717-4277-b739-393ecb2081e4","Type":"ContainerStarted","Data":"963b01f6bff28045993d91540cfc52ed8041ca3f8e4835d3d714afe976826976"} Dec 03 16:58:01 crc kubenswrapper[4758]: W1203 16:58:01.315180 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6da11ac_6aa4_43d3_927e_315d72618eb6.slice/crio-0a1752b390e8e3826595e6df998c8d6733b58528c5c158308fca5a8949c21e07 WatchSource:0}: Error finding container 0a1752b390e8e3826595e6df998c8d6733b58528c5c158308fca5a8949c21e07: Status 404 returned error can't find the container with id 0a1752b390e8e3826595e6df998c8d6733b58528c5c158308fca5a8949c21e07 Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.316627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" event={"ID":"014604aa-7ce2-48ed-b924-00ac8db7abb1","Type":"ContainerStarted","Data":"36c4d94b3a73cf20e1abdd0dc584ed1592ef562c9dc7f15f8b95f0ce1bf320f9"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.321364 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" event={"ID":"813a5ac2-3ab0-4b77-9017-f79f849e7ce1","Type":"ContainerStarted","Data":"5104f78f250aa464fddc245929e382139bf1875f6d4d416e22d4f62750fc7d10"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.327712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" event={"ID":"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2","Type":"ContainerStarted","Data":"e9228761696e797570e684d6e5bd6f9a82e8388e7447decd59ed27b5dc601049"} Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.328389 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-qkmml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.328458 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qkmml" podUID="ceabfcec-18d7-4f41-af6c-b31edfcdbe80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.374027 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.375072 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.875028518 +0000 UTC m=+137.076405379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.377543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.379108 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.879083577 +0000 UTC m=+137.080460438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.481135 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.481513 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.981475492 +0000 UTC m=+137.182852353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.482058 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.482581 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:01.982559232 +0000 UTC m=+137.183936093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.505690 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" podStartSLOduration=118.50562644 podStartE2EDuration="1m58.50562644s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:01.49329118 +0000 UTC m=+136.694668041" watchObservedRunningTime="2025-12-03 16:58:01.50562644 +0000 UTC m=+136.707003301" Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.569867 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4f6xq"] Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.587311 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.587587 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.087561458 +0000 UTC m=+137.288938329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.587719 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.588075 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.088068321 +0000 UTC m=+137.289445182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.600981 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" podStartSLOduration=118.600962746 podStartE2EDuration="1m58.600962746s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:01.600251178 +0000 UTC m=+136.801628039" watchObservedRunningTime="2025-12-03 16:58:01.600962746 +0000 UTC m=+136.802339607" Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.618047 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gsnzz" podStartSLOduration=118.618022695 podStartE2EDuration="1m58.618022695s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:01.616805722 +0000 UTC m=+136.818182593" watchObservedRunningTime="2025-12-03 16:58:01.618022695 +0000 UTC m=+136.819399556" Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.688910 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.689124 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.189058559 +0000 UTC m=+137.390435420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.689364 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.689924 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.189916062 +0000 UTC m=+137.391292923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.790156 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.790630 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.290610693 +0000 UTC m=+137.491987554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.894373 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.895170 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.395146546 +0000 UTC m=+137.596523407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:01 crc kubenswrapper[4758]: I1203 16:58:01.998501 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:01 crc kubenswrapper[4758]: E1203 16:58:01.998993 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.49897172 +0000 UTC m=+137.700348581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.101196 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.101659 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.601644073 +0000 UTC m=+137.803020934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.202444 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.202689 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.702648202 +0000 UTC m=+137.904025063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.203652 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.204144 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.704132261 +0000 UTC m=+137.905509122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.311848 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.312514 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.812491568 +0000 UTC m=+138.013868429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.414792 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.415721 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:02.915705065 +0000 UTC m=+138.117081926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.463500 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" event={"ID":"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b","Type":"ContainerStarted","Data":"06b66198fb62bf786ff3626b246f10276777a81c186a1e1d13d9e890fc1b44da"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.481074 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" event={"ID":"10048458-f2d3-4599-a7d2-727fe81b6006","Type":"ContainerStarted","Data":"766918485b917ae4a37729f8b07a82589ef9421d022372e93a0f148d70d11e78"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.497256 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" event={"ID":"f0dd0c9a-f13a-4d9b-b812-da766c97e831","Type":"ContainerStarted","Data":"6b3461b77d601779cb2b119662397062bf6f2e957cf604c484fb2a2e22f9f705"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.503919 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hvg6p" event={"ID":"10a66e2c-cb51-4eec-864b-d8288ca3e281","Type":"ContainerStarted","Data":"35f77fb9732af1293554269d31bd53d0d4e381ef9719257f969a46f3654d43c6"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.516329 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.516935 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.01690984 +0000 UTC m=+138.218286701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.519412 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.519857 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.019840208 +0000 UTC m=+138.221217069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.544481 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hqwcv" podStartSLOduration=119.544456768 podStartE2EDuration="1m59.544456768s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:02.543235335 +0000 UTC m=+137.744612196" watchObservedRunningTime="2025-12-03 16:58:02.544456768 +0000 UTC m=+137.745833629" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.579121 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8dmlt" event={"ID":"a35fa7d5-51a2-4e11-b805-e79d0ddbafa2","Type":"ContainerStarted","Data":"2236fcb0f2038081f19ffe33e479870876a528e355a402e1e272d35e6b9b2143"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.602418 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hvg6p" podStartSLOduration=6.6023927019999995 podStartE2EDuration="6.602392702s" podCreationTimestamp="2025-12-03 16:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:02.602097053 +0000 UTC m=+137.803473914" watchObservedRunningTime="2025-12-03 16:58:02.602392702 +0000 UTC m=+137.803769563" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.638224 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.639773 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.139740363 +0000 UTC m=+138.341117374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.670010 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" event={"ID":"ecdc8ba7-41d8-419a-ac31-52278e353468","Type":"ContainerStarted","Data":"5faaa8a2ecd05ac981f1028af00a2e0016f5c1a47f6542fde38cd9523d437592"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.721568 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-smpbk" event={"ID":"fd8026a4-bc63-4912-94a8-897cf3c0bea6","Type":"ContainerStarted","Data":"ded6e221f61e8d90eada4d0d6257ec249959519253e9d1ac120748cc11e3845c"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.746166 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.750324 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.250288857 +0000 UTC m=+138.451665718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.775892 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" event={"ID":"c138db90-10f2-4e2b-8cb3-8b368cbc1a7c","Type":"ContainerStarted","Data":"d0fbd197a49e5735c049243f517dae6aba3f908ad1db2298b12b8e935db9dfdd"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.788404 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.820106 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" podStartSLOduration=119.820080409 podStartE2EDuration="1m59.820080409s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:02.819218556 +0000 UTC m=+138.020595417" watchObservedRunningTime="2025-12-03 16:58:02.820080409 +0000 UTC m=+138.021457270" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.820370 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gz4q6" podStartSLOduration=119.820366297 podStartE2EDuration="1m59.820366297s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:02.708364033 +0000 UTC m=+137.909740894" watchObservedRunningTime="2025-12-03 16:58:02.820366297 +0000 UTC m=+138.021743158" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.857609 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.858101 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.358062188 +0000 UTC m=+138.559439049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.892877 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" event={"ID":"71fd30bb-8da6-4b86-8dd6-56274220a60b","Type":"ContainerStarted","Data":"358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.897175 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.900343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" event={"ID":"813a5ac2-3ab0-4b77-9017-f79f849e7ce1","Type":"ContainerStarted","Data":"7e976b1d7cf589ab19138662f12407f1065a571c5568c35c98268477efcceb05"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.930280 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.950695 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4l6l4" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.957192 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" event={"ID":"cd60f8c1-b60d-4d72-b25d-1496429ebcae","Type":"ContainerStarted","Data":"3a0f0f9b3f66f333ce3e71e76a773609b8355d3a168d1d55e857f3307b0b59a6"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.962369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.965699 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" event={"ID":"36d8e040-d731-4f82-ba5b-ace1145e15f6","Type":"ContainerStarted","Data":"1629d830aff98992bc1ea3658156f4e6454ffda13b753ad6e4bc84ef45741e4f"} Dec 03 16:58:02 crc kubenswrapper[4758]: E1203 16:58:02.965922 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.465901579 +0000 UTC m=+138.667278440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.968890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c9b57" event={"ID":"a6da11ac-6aa4-43d3-927e-315d72618eb6","Type":"ContainerStarted","Data":"0a1752b390e8e3826595e6df998c8d6733b58528c5c158308fca5a8949c21e07"} Dec 03 16:58:02 crc kubenswrapper[4758]: I1203 16:58:02.992526 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dcg47" event={"ID":"67bbd99c-59f5-4bb0-af30-d503cb341f2c","Type":"ContainerStarted","Data":"29a603ab18dcd3ac864bdd1d84af46f02476e96d42b22c047c379cac234d59d7"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.012480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kzv6z" event={"ID":"a32f32fc-e1aa-402c-93ec-f73446955bd0","Type":"ContainerStarted","Data":"7a2835a436f3300583867ea64cb21b13fed5b722bdd540d7a22e4e65d4355463"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.012530 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kzv6z" event={"ID":"a32f32fc-e1aa-402c-93ec-f73446955bd0","Type":"ContainerStarted","Data":"3d18e8d674c5a8ab06381aa52458283e147d63ef2bb4b4301bee2b2160ccee05"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.015973 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" event={"ID":"32009def-8f0b-4b71-961c-422ccb9aafea","Type":"ContainerStarted","Data":"4f0e5cf45d00b136848306afd759074fc34ded0aa59ef8ed1edb3eb6f3fa864e"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.016861 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.018530 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" podStartSLOduration=120.01849558 podStartE2EDuration="2m0.01849558s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:02.978481337 +0000 UTC m=+138.179858198" watchObservedRunningTime="2025-12-03 16:58:03.01849558 +0000 UTC m=+138.219872431" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.021005 4758 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t4xg7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.021055 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" podUID="32009def-8f0b-4b71-961c-422ccb9aafea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.035783 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jxqhr" event={"ID":"58900711-350a-4753-888d-7a6ff520abd1","Type":"ContainerStarted","Data":"c3d81fbf10c0284443bb52cc94a266810abf475eb4b9149b6b23a3262cb5570f"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.052020 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" event={"ID":"784b6e22-0844-494f-ac12-2fa16fccc21b","Type":"ContainerStarted","Data":"8463cb4e7d31b9e2849d853142b281ab787abb0a49bbe46c26170a6dfaee142c"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.063220 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.065091 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.565075179 +0000 UTC m=+138.766452040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.067171 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jmjcs" podStartSLOduration=120.067147465 podStartE2EDuration="2m0.067147465s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.065619973 +0000 UTC m=+138.266996834" watchObservedRunningTime="2025-12-03 16:58:03.067147465 +0000 UTC m=+138.268524326" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.097969 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" event={"ID":"74a07ef0-49b8-442a-b011-fc80b710cb80","Type":"ContainerStarted","Data":"32b99358cf844e3065734216670c98a4f67b725198b271ed0887cb23361756c2"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.119582 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-dcg47" podStartSLOduration=7.11955395 podStartE2EDuration="7.11955395s" podCreationTimestamp="2025-12-03 16:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.119339514 +0000 UTC m=+138.320716385" watchObservedRunningTime="2025-12-03 16:58:03.11955395 +0000 UTC m=+138.320930811" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.166819 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.183022 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.682995891 +0000 UTC m=+138.884372752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.271449 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.273454 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.773426026 +0000 UTC m=+138.974802887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.309234 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r8sqq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.309346 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.359823 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.359893 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" event={"ID":"acb4a12e-76ef-4e9a-b6e4-18e4fe4ff319","Type":"ContainerStarted","Data":"428fd6d1300220887713e405d20dffc2ac4ee026f5e74386db0be69d60287dfd"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.359939 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" event={"ID":"abe3b323-2c4b-42a0-9357-a29866107e8a","Type":"ContainerStarted","Data":"bcea9e782347f72e0cc4e8f30725a827a7c28dd70b4c9545f7b2a77aab1e28c1"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.359959 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.359975 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" event={"ID":"dcbd62f3-c933-4182-8172-3ef2fc3220ec","Type":"ContainerStarted","Data":"2ae3bd37254f452d9060dbbf23689665fff5d5e4af2a4a3d895b34241dcc72e4"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.359988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" event={"ID":"83593597-7147-4b61-be7c-e05b165bdebe","Type":"ContainerStarted","Data":"128fc95a96bc235f23f52262de8749879b7e5255b470a20ef23180daf4dcfdd8"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.372162 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" event={"ID":"bf1a9d85-21f7-446e-9ca3-e2d3d9010130","Type":"ContainerStarted","Data":"978b3936c11876fc2b6d573465c9349b74018a340d243f46e51eccc1b1ee4148"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.377571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.378040 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.878026062 +0000 UTC m=+139.079402923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.401616 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" event={"ID":"b6b9a502-ea7d-4dae-9539-89b710c40171","Type":"ContainerStarted","Data":"f59da3ac935c59d8df9d2e9b114b6f173aec456cf34f691722f21f89305949ae"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.402049 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" event={"ID":"b6b9a502-ea7d-4dae-9539-89b710c40171","Type":"ContainerStarted","Data":"f5139840ca9206878cdb223b1c993116347edafce990dd181bc965afbcc4d923"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.403450 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.430442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" event={"ID":"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8","Type":"ContainerStarted","Data":"39d4fa446bc32dd8748a4885717bd23bdb9137836bb606e4084838acbd9feeb2"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.441041 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" event={"ID":"92c7bc44-a787-4fab-8213-e316dc8ca4dc","Type":"ContainerStarted","Data":"3f9c932b9520a4d41370521851a908c5201f4cbce948e65a8ba504c1a1b95c5a"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.442182 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.463210 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.476445 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" event={"ID":"c30fe471-6fb2-467a-b63e-2fc8f915c948","Type":"ContainerStarted","Data":"97c83131a914a6b8fcba7cf698bed9360d69706ddbf60f03419e618bc55482f2"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.476510 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" event={"ID":"c30fe471-6fb2-467a-b63e-2fc8f915c948","Type":"ContainerStarted","Data":"668a66256a32352e1ce33cafacd211bd718d0255aea36e696887b4b2f9c33cdb"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.477530 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.478515 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.480295 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:03.980262003 +0000 UTC m=+139.181639054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.505979 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kzv6z" podStartSLOduration=120.505931141 podStartE2EDuration="2m0.505931141s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.443988481 +0000 UTC m=+138.645365342" watchObservedRunningTime="2025-12-03 16:58:03.505931141 +0000 UTC m=+138.707307992" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.509297 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.528725 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.551588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" event={"ID":"93b3d24c-56a8-485c-820d-7567ed1a6733","Type":"ContainerStarted","Data":"528f83fb7b20a4a9c5e80241c112a89a548f7fddaaad4a05ef36149fe85fcc40"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.551655 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" event={"ID":"93b3d24c-56a8-485c-820d-7567ed1a6733","Type":"ContainerStarted","Data":"9d31c724ff42f74e565768f856c392640a4072082fc7ddd4ef4ad550a3c4cbe8"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.600828 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jxqhr" podStartSLOduration=120.553659821 podStartE2EDuration="2m0.553659821s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.50885322 +0000 UTC m=+138.710230081" watchObservedRunningTime="2025-12-03 16:58:03.553659821 +0000 UTC m=+138.755036672" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.601548 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" podStartSLOduration=119.601517764 podStartE2EDuration="1m59.601517764s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.551554555 +0000 UTC m=+138.752931416" watchObservedRunningTime="2025-12-03 16:58:03.601517764 +0000 UTC m=+138.802894625" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.603823 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.604250 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.104232717 +0000 UTC m=+139.305609578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.628066 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lxkjl" podStartSLOduration=119.628031935 podStartE2EDuration="1m59.628031935s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.61851009 +0000 UTC m=+138.819886961" watchObservedRunningTime="2025-12-03 16:58:03.628031935 +0000 UTC m=+138.829408796" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.633468 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" event={"ID":"f006d39c-146c-4461-aaa0-a81a6b7d535a","Type":"ContainerStarted","Data":"c89070803b44fd7c2f596ec17503db6247e4332187be16e2ce39900ce6433434"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.648578 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" event={"ID":"bd7d3310-f1bb-4557-a852-bd4f1f72eb60","Type":"ContainerStarted","Data":"795f3d1b19e1bac14502bf7d1ace200fe7f4e14b4b72f4da92e70745201ffe01"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.693063 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" event={"ID":"82f6d6a9-f455-4dc5-bdfb-ac3616141bad","Type":"ContainerStarted","Data":"4002fb511aeb1fccc69dba4937fa0615fba8e38ad21eba48d5941fd7c35e9ede"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.700647 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" event={"ID":"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63","Type":"ContainerStarted","Data":"484604c9fb923340283cd13971b693cae8bf1e2ff5f689509e61e95530284626"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.700703 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" event={"ID":"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63","Type":"ContainerStarted","Data":"a6910196e38744f39586e5a862d74ed235d8fc6ea99595b8aec70cd1c686c8f7"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.704078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" event={"ID":"886f030a-c717-4277-b739-393ecb2081e4","Type":"ContainerStarted","Data":"cb5f1783808962262d7e213cf2a4d8c33df2669539a9fb25dc259fd2cc351b56"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.704319 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.704525 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.204498836 +0000 UTC m=+139.405875697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.728953 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" event={"ID":"c3615454-def8-4637-884b-a4c96fe468d6","Type":"ContainerStarted","Data":"809a86eb7e2a478ff7b0d35e53ea2d05d7ff8fd089b7666ebbfaceedf2f73ac5"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.753723 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" podStartSLOduration=119.753676735 podStartE2EDuration="1m59.753676735s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.730345269 +0000 UTC m=+138.931722130" watchObservedRunningTime="2025-12-03 16:58:03.753676735 +0000 UTC m=+138.955053596" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.758206 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" event={"ID":"40add3ff-6f8a-4b7d-a385-454665ca770c","Type":"ContainerStarted","Data":"fb185e2529f908dbc284c014253f588d5747f794906686a455ca35b29bdf001c"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.768800 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" podStartSLOduration=119.76877599 podStartE2EDuration="1m59.76877599s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.767721912 +0000 UTC m=+138.969098773" watchObservedRunningTime="2025-12-03 16:58:03.76877599 +0000 UTC m=+138.970152851" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.792532 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" event={"ID":"71308db0-b934-4c92-b90c-aaf1c6b1b41c","Type":"ContainerStarted","Data":"cc50c38f4bc54d617e83b3d0384acdfc21f0d4f922adc1a7eb60e6ed7c3e279e"} Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.792587 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.797702 4758 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4f6xq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.797762 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" podUID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.801775 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.808521 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.809138 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.309122201 +0000 UTC m=+139.510499062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.818036 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:03 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:03 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:03 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.818115 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.889533 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljss6" podStartSLOduration=120.889513658 podStartE2EDuration="2m0.889513658s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.887673679 +0000 UTC m=+139.089050540" watchObservedRunningTime="2025-12-03 16:58:03.889513658 +0000 UTC m=+139.090890519" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.890353 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" podStartSLOduration=119.890348721 podStartE2EDuration="1m59.890348721s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.822008408 +0000 UTC m=+139.023385269" watchObservedRunningTime="2025-12-03 16:58:03.890348721 +0000 UTC m=+139.091725582" Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.914775 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:03 crc kubenswrapper[4758]: E1203 16:58:03.916572 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.416548652 +0000 UTC m=+139.617925513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:03 crc kubenswrapper[4758]: I1203 16:58:03.942916 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" podStartSLOduration=119.942892399 podStartE2EDuration="1m59.942892399s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:03.9410551 +0000 UTC m=+139.142431961" watchObservedRunningTime="2025-12-03 16:58:03.942892399 +0000 UTC m=+139.144269260" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.016491 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.016884 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.516869083 +0000 UTC m=+139.718245934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.105524 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" podStartSLOduration=121.1055012 podStartE2EDuration="2m1.1055012s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.038863483 +0000 UTC m=+139.240240344" watchObservedRunningTime="2025-12-03 16:58:04.1055012 +0000 UTC m=+139.306878061" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.124696 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.125152 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.625127746 +0000 UTC m=+139.826504607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.180344 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" podStartSLOduration=120.180320266 podStartE2EDuration="2m0.180320266s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.179585526 +0000 UTC m=+139.380962387" watchObservedRunningTime="2025-12-03 16:58:04.180320266 +0000 UTC m=+139.381697127" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.181356 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" podStartSLOduration=120.181348913 podStartE2EDuration="2m0.181348913s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.152609403 +0000 UTC m=+139.353986254" watchObservedRunningTime="2025-12-03 16:58:04.181348913 +0000 UTC m=+139.382725774" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.230166 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.230621 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.730605065 +0000 UTC m=+139.931981926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.274208 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" podStartSLOduration=121.274182833 podStartE2EDuration="2m1.274182833s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.273184776 +0000 UTC m=+139.474561637" watchObservedRunningTime="2025-12-03 16:58:04.274182833 +0000 UTC m=+139.475559694" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.276454 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" podStartSLOduration=120.276442033 podStartE2EDuration="2m0.276442033s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.237573062 +0000 UTC m=+139.438949923" watchObservedRunningTime="2025-12-03 16:58:04.276442033 +0000 UTC m=+139.477818894" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.332809 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.333138 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.833085053 +0000 UTC m=+140.034461914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.355323 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" podStartSLOduration=120.355301669 podStartE2EDuration="2m0.355301669s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.309737037 +0000 UTC m=+139.511113898" watchObservedRunningTime="2025-12-03 16:58:04.355301669 +0000 UTC m=+139.556678530" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.438697 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.439106 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:04.939090085 +0000 UTC m=+140.140466946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.540136 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.540376 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.040340971 +0000 UTC m=+140.241717832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.540691 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.541118 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.04107685 +0000 UTC m=+140.242453711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.642491 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.643151 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.143110686 +0000 UTC m=+140.344487697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.744571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.745088 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.245066411 +0000 UTC m=+140.446443262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.806798 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:04 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:04 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:04 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.806864 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.835278 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xgmkj" event={"ID":"82f6d6a9-f455-4dc5-bdfb-ac3616141bad","Type":"ContainerStarted","Data":"85ea23acbaf583b0f8fdb00ee40af81a156e4d245fe8ab0c32d605d7b86d7b2c"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.840204 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" event={"ID":"83593597-7147-4b61-be7c-e05b165bdebe","Type":"ContainerStarted","Data":"531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.841090 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r8sqq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.841137 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.845262 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.845437 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.345408062 +0000 UTC m=+140.546784923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.847336 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.848932 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.348920665 +0000 UTC m=+140.550297526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.849661 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" event={"ID":"40add3ff-6f8a-4b7d-a385-454665ca770c","Type":"ContainerStarted","Data":"e35147313938fe9f0e4a8e31300a6346d50134d58acbad1c48448892fc870251"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.878279 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" event={"ID":"784b6e22-0844-494f-ac12-2fa16fccc21b","Type":"ContainerStarted","Data":"d6fe24afdcc8552cdd29004ce117077ae3a1ee09ae43159edfe95d5aacaeeb20"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.878350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" event={"ID":"784b6e22-0844-494f-ac12-2fa16fccc21b","Type":"ContainerStarted","Data":"df5925768bb91dde1ab33c02cfa4ee66d4191dcc0642208995b509676486f8db"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.879120 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.893902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" event={"ID":"014604aa-7ce2-48ed-b924-00ac8db7abb1","Type":"ContainerStarted","Data":"2c75d76f63b40274494e0524f615b56a1bd9f287909aa0f41d0098119ad6a3b3"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.899583 4758 generic.go:334] "Generic (PLEG): container finished" podID="dcbd62f3-c933-4182-8172-3ef2fc3220ec" containerID="07925e0a0bec0f3bd36d98fbe2a44a4f43fe4cab788d9f55d5559ee351e18313" exitCode=0 Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.899966 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" event={"ID":"dcbd62f3-c933-4182-8172-3ef2fc3220ec","Type":"ContainerStarted","Data":"b8733e35c00ce11abf919fac6a9bf76f16917cd3715ddb982d04a80fc0c0cb00"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.900032 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" event={"ID":"dcbd62f3-c933-4182-8172-3ef2fc3220ec","Type":"ContainerDied","Data":"07925e0a0bec0f3bd36d98fbe2a44a4f43fe4cab788d9f55d5559ee351e18313"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.900157 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.905973 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5hdt9" event={"ID":"bd7d3310-f1bb-4557-a852-bd4f1f72eb60","Type":"ContainerStarted","Data":"fbbc8213af2d184035a02cc0b00f0da319473ad25fed1357492fcf6a407fc267"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.910158 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hv699" podStartSLOduration=121.910137287 podStartE2EDuration="2m1.910137287s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.909224433 +0000 UTC m=+140.110601284" watchObservedRunningTime="2025-12-03 16:58:04.910137287 +0000 UTC m=+140.111514138" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.911708 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" podStartSLOduration=121.911701299 podStartE2EDuration="2m1.911701299s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:04.361727641 +0000 UTC m=+139.563104502" watchObservedRunningTime="2025-12-03 16:58:04.911701299 +0000 UTC m=+140.113078160" Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.928759 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" event={"ID":"f006d39c-146c-4461-aaa0-a81a6b7d535a","Type":"ContainerStarted","Data":"09009ce9edeee347ca1185092b2548a89741ab9257ec5b74c446970062f45550"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.928846 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" event={"ID":"f006d39c-146c-4461-aaa0-a81a6b7d535a","Type":"ContainerStarted","Data":"514655e75bcbe22bcdd8409cafcf18fd5cfce6235b23f2d6797d502ef626e810"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.949351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:04 crc kubenswrapper[4758]: E1203 16:58:04.951084 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.451032484 +0000 UTC m=+140.652409345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.969459 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c9b57" event={"ID":"a6da11ac-6aa4-43d3-927e-315d72618eb6","Type":"ContainerStarted","Data":"c95d7815c8a0dc7291ed53f1032151d6ee2b65ab55e4ad82108ace9db9786c32"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.969982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c9b57" event={"ID":"a6da11ac-6aa4-43d3-927e-315d72618eb6","Type":"ContainerStarted","Data":"0621a82ad70259b28d4db8de896b7aa03d0e5f468f328a3a94676d1a617c5b03"} Dec 03 16:58:04 crc kubenswrapper[4758]: I1203 16:58:04.970842 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-c9b57" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.001407 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" podStartSLOduration=121.001388174 podStartE2EDuration="2m1.001388174s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.000001247 +0000 UTC m=+140.201378118" watchObservedRunningTime="2025-12-03 16:58:05.001388174 +0000 UTC m=+140.202765035" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.020035 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" event={"ID":"36d8e040-d731-4f82-ba5b-ace1145e15f6","Type":"ContainerStarted","Data":"9aa6db690f485b30405ff428fdd982b93dfce2a5cd672e03016428b7b8d4eea9"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.035021 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" event={"ID":"71308db0-b934-4c92-b90c-aaf1c6b1b41c","Type":"ContainerStarted","Data":"533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.046498 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xlvqm" event={"ID":"bf1a9d85-21f7-446e-9ca3-e2d3d9010130","Type":"ContainerStarted","Data":"ef3cde7a4dcd20ec1cb1686fe6b960a2bc1c1723ff463509248ff7c4a552c702"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.053216 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.056481 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" podStartSLOduration=122.056463491 podStartE2EDuration="2m2.056463491s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.056166353 +0000 UTC m=+140.257543214" watchObservedRunningTime="2025-12-03 16:58:05.056463491 +0000 UTC m=+140.257840352" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.057317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.064113 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.564094586 +0000 UTC m=+140.765471447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.080116 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x976j" event={"ID":"7a8ccf1a-e3c0-4f87-8e92-69b7d15cca63","Type":"ContainerStarted","Data":"44b70a287d6b360d57fd561d4ac2f47621f83c5403c586b9665d964085ff48d4"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.101410 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7lnzm" podStartSLOduration=121.101383415 podStartE2EDuration="2m1.101383415s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.100357098 +0000 UTC m=+140.301733959" watchObservedRunningTime="2025-12-03 16:58:05.101383415 +0000 UTC m=+140.302760266" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.146535 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xj7gk" event={"ID":"92c7bc44-a787-4fab-8213-e316dc8ca4dc","Type":"ContainerStarted","Data":"2118dc191027082c22ee44ce45bd39b0c42dd64469290617e4dad90e63d5948d"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.157096 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" event={"ID":"32009def-8f0b-4b71-961c-422ccb9aafea","Type":"ContainerStarted","Data":"e14ebca98c4a99b94df4ca8ba43b4a1c195f5965de27d76b62b965ae147512f7"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.164492 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.165504 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.665470784 +0000 UTC m=+140.866847645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.179341 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" event={"ID":"93b3d24c-56a8-485c-820d-7567ed1a6733","Type":"ContainerStarted","Data":"7ee23ff905dc5e52acb0a017adb45e921f627c30da0c0c77b0b8d833862c0896"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.212449 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-nnc2j" podStartSLOduration=121.212427124 podStartE2EDuration="2m1.212427124s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.208296453 +0000 UTC m=+140.409673314" watchObservedRunningTime="2025-12-03 16:58:05.212427124 +0000 UTC m=+140.413803985" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.227317 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" event={"ID":"e9ea76f4-50d8-416a-85dd-897f08112fa3","Type":"ContainerStarted","Data":"4b49759c492bb9575f624e2fa50da742843db316e447bdd97293f20218ea62ed"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.247094 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntv4j" event={"ID":"abe3b323-2c4b-42a0-9357-a29866107e8a","Type":"ContainerStarted","Data":"cffdde18483b2d163382c95d22c0a112f67e2d2044caa7fda367cef9d177588b"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.266899 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.267531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" event={"ID":"c3615454-def8-4637-884b-a4c96fe468d6","Type":"ContainerStarted","Data":"be0d4204aed83354760e0a1aa057876cd4aebb6c935ec01a31414e1c3a809fcf"} Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.269979 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.769965007 +0000 UTC m=+140.971341868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.277769 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6r465" podStartSLOduration=122.277753325 podStartE2EDuration="2m2.277753325s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.277597701 +0000 UTC m=+140.478974562" watchObservedRunningTime="2025-12-03 16:58:05.277753325 +0000 UTC m=+140.479130186" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.279819 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-c9b57" podStartSLOduration=9.279812651 podStartE2EDuration="9.279812651s" podCreationTimestamp="2025-12-03 16:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.246723013 +0000 UTC m=+140.448099874" watchObservedRunningTime="2025-12-03 16:58:05.279812651 +0000 UTC m=+140.481189512" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.291270 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" event={"ID":"f4a6fd6b-4562-42ba-85d6-710a8f6bb98b","Type":"ContainerStarted","Data":"423a12f089b073f23751488a91bd3730a0aeea6217812cb39102f665e2d886a7"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.323935 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" event={"ID":"10048458-f2d3-4599-a7d2-727fe81b6006","Type":"ContainerStarted","Data":"ebd6a9b33c5f92667b9b547c4ebd8dcb918d4495ab3a597be3e354d3227ca6f8"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.324317 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" event={"ID":"10048458-f2d3-4599-a7d2-727fe81b6006","Type":"ContainerStarted","Data":"ca8ae16a24cea43d3098912ac67cc4cac885758b3232e50c48d9083aff245ce4"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.324783 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwfxl" podStartSLOduration=122.324757176 podStartE2EDuration="2m2.324757176s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.323980685 +0000 UTC m=+140.525357546" watchObservedRunningTime="2025-12-03 16:58:05.324757176 +0000 UTC m=+140.526134037" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.334879 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7kgk7"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.336065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.349319 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nx6lj" event={"ID":"fbb99ef5-6a95-43f4-a404-1d4a14cf28e8","Type":"ContainerStarted","Data":"1d4a37eb2a3bbb13903facd5391c8c1fb4df54b04bcc7746f599245427feb946"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.358558 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.366882 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" event={"ID":"cd60f8c1-b60d-4d72-b25d-1496429ebcae","Type":"ContainerStarted","Data":"049f201859101dfef2de3033e90f454cd401fd622560a746f9fb6e726f7996c8"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.367470 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.369308 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.86929073 +0000 UTC m=+141.070667591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.391599 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" event={"ID":"74a07ef0-49b8-442a-b011-fc80b710cb80","Type":"ContainerStarted","Data":"ec978dc90f376c41b0639420f83ee6869ebff8a5e0d332aeb85feec685f3804e"} Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.393447 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7kgk7"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.432062 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" podStartSLOduration=121.432041713 podStartE2EDuration="2m1.432041713s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.430986815 +0000 UTC m=+140.632363676" watchObservedRunningTime="2025-12-03 16:58:05.432041713 +0000 UTC m=+140.633418574" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.473486 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-utilities\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.473572 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnbjk\" (UniqueName: \"kubernetes.io/projected/f673f32d-f81f-4fe9-ab86-801fe50a46da-kube-api-access-dnbjk\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.473632 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-catalog-content\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.473732 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.478997 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t4xg7" Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.482428 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:05.982402554 +0000 UTC m=+141.183779415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.540078 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d4b9m"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.541185 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.553378 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.569779 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4b9m"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.571353 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-qjwkt" podStartSLOduration=122.571326738 podStartE2EDuration="2m2.571326738s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.568009969 +0000 UTC m=+140.769386830" watchObservedRunningTime="2025-12-03 16:58:05.571326738 +0000 UTC m=+140.772703599" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.575184 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.576079 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-utilities\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.576418 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnbjk\" (UniqueName: \"kubernetes.io/projected/f673f32d-f81f-4fe9-ab86-801fe50a46da-kube-api-access-dnbjk\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.576622 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-catalog-content\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.577605 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.077572896 +0000 UTC m=+141.278949757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.579242 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-catalog-content\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.582013 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-utilities\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.623290 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" podStartSLOduration=122.623273501 podStartE2EDuration="2m2.623273501s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.621065202 +0000 UTC m=+140.822442063" watchObservedRunningTime="2025-12-03 16:58:05.623273501 +0000 UTC m=+140.824650362" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.647232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnbjk\" (UniqueName: \"kubernetes.io/projected/f673f32d-f81f-4fe9-ab86-801fe50a46da-kube-api-access-dnbjk\") pod \"community-operators-7kgk7\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.678294 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-utilities\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.678387 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.678460 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-catalog-content\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.678513 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms74c\" (UniqueName: \"kubernetes.io/projected/b43b0669-6da9-438d-bc36-90456cc2611c-kube-api-access-ms74c\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.678925 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.178911763 +0000 UTC m=+141.380288624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.686518 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.725156 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-srplx" podStartSLOduration=121.725126552 podStartE2EDuration="2m1.725126552s" podCreationTimestamp="2025-12-03 16:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.665055241 +0000 UTC m=+140.866432102" watchObservedRunningTime="2025-12-03 16:58:05.725126552 +0000 UTC m=+140.926503413" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.741220 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wm6jh" podStartSLOduration=122.741188493 podStartE2EDuration="2m2.741188493s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:05.731255276 +0000 UTC m=+140.932632137" watchObservedRunningTime="2025-12-03 16:58:05.741188493 +0000 UTC m=+140.942565344" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.765747 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l55d4"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.766997 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.779928 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.780312 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-catalog-content\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.780472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms74c\" (UniqueName: \"kubernetes.io/projected/b43b0669-6da9-438d-bc36-90456cc2611c-kube-api-access-ms74c\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.780586 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-utilities\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.781289 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-utilities\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.782084 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.282050479 +0000 UTC m=+141.483427330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.782611 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-catalog-content\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.793939 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l55d4"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.826228 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms74c\" (UniqueName: \"kubernetes.io/projected/b43b0669-6da9-438d-bc36-90456cc2611c-kube-api-access-ms74c\") pod \"certified-operators-d4b9m\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.831490 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:05 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:05 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:05 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.832161 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.869205 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.889742 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thv2f\" (UniqueName: \"kubernetes.io/projected/8dc530f3-6688-414d-9f8a-bb191ad5a264-kube-api-access-thv2f\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.890082 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-catalog-content\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.890197 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.890278 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-utilities\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:05 crc kubenswrapper[4758]: E1203 16:58:05.890692 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.390663331 +0000 UTC m=+141.592040192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.921766 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jm64k"] Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.922982 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:05 crc kubenswrapper[4758]: I1203 16:58:05.938202 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jm64k"] Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.044988 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.544951099 +0000 UTC m=+141.746327960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.045485 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046195 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-catalog-content\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-utilities\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046487 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046586 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-utilities\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046753 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92f86\" (UniqueName: \"kubernetes.io/projected/db60029f-ec45-4c51-98cd-ca184c65c38d-kube-api-access-92f86\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046896 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thv2f\" (UniqueName: \"kubernetes.io/projected/8dc530f3-6688-414d-9f8a-bb191ad5a264-kube-api-access-thv2f\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.046986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-catalog-content\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.047591 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-catalog-content\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.048142 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-utilities\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.048509 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.548491254 +0000 UTC m=+141.749868115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.105366 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thv2f\" (UniqueName: \"kubernetes.io/projected/8dc530f3-6688-414d-9f8a-bb191ad5a264-kube-api-access-thv2f\") pod \"community-operators-l55d4\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.150503 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.151278 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92f86\" (UniqueName: \"kubernetes.io/projected/db60029f-ec45-4c51-98cd-ca184c65c38d-kube-api-access-92f86\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.151320 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-catalog-content\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.151380 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-utilities\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.151866 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-utilities\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.151962 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.651939388 +0000 UTC m=+141.853316249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.152502 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-catalog-content\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.234070 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92f86\" (UniqueName: \"kubernetes.io/projected/db60029f-ec45-4c51-98cd-ca184c65c38d-kube-api-access-92f86\") pod \"certified-operators-jm64k\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.253152 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.253530 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.753518162 +0000 UTC m=+141.954895023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.280204 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.355813 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.356223 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.856206376 +0000 UTC m=+142.057583237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.402134 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.457618 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.461610 4758 generic.go:334] "Generic (PLEG): container finished" podID="886f030a-c717-4277-b739-393ecb2081e4" containerID="cb5f1783808962262d7e213cf2a4d8c33df2669539a9fb25dc259fd2cc351b56" exitCode=0 Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.461718 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" event={"ID":"886f030a-c717-4277-b739-393ecb2081e4","Type":"ContainerDied","Data":"cb5f1783808962262d7e213cf2a4d8c33df2669539a9fb25dc259fd2cc351b56"} Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.464073 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:06.964037758 +0000 UTC m=+142.165414619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.518485 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" event={"ID":"c3615454-def8-4637-884b-a4c96fe468d6","Type":"ContainerStarted","Data":"67f60733f84b50a399467ce15dfc4ff2a617e8ab4883f7961fd18c206f4626f5"} Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.522546 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r8sqq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.522707 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.570326 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.572719 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.07266752 +0000 UTC m=+142.274044381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.693243 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.694156 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.194134388 +0000 UTC m=+142.395511249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.796273 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.796510 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.296493503 +0000 UTC m=+142.497870364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.806230 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:06 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:06 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:06 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.806289 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.842521 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7kgk7"] Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.862221 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4b9m"] Dec 03 16:58:06 crc kubenswrapper[4758]: W1203 16:58:06.922192 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb43b0669_6da9_438d_bc36_90456cc2611c.slice/crio-a87273ce688f26abfd57001490d4479b2df9df8458882a8455945015b385b72a WatchSource:0}: Error finding container a87273ce688f26abfd57001490d4479b2df9df8458882a8455945015b385b72a: Status 404 returned error can't find the container with id a87273ce688f26abfd57001490d4479b2df9df8458882a8455945015b385b72a Dec 03 16:58:06 crc kubenswrapper[4758]: I1203 16:58:06.924944 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:06 crc kubenswrapper[4758]: E1203 16:58:06.925444 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.42542275 +0000 UTC m=+142.626799601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.028468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:07 crc kubenswrapper[4758]: E1203 16:58:07.029061 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.529034829 +0000 UTC m=+142.730411690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.112316 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l55d4"] Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.130975 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:07 crc kubenswrapper[4758]: E1203 16:58:07.131463 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.631440125 +0000 UTC m=+142.832816986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.159496 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jm64k"] Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.232175 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:07 crc kubenswrapper[4758]: E1203 16:58:07.232388 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.73234936 +0000 UTC m=+142.933726211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.232974 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:07 crc kubenswrapper[4758]: E1203 16:58:07.233414 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.733393189 +0000 UTC m=+142.934770050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94vfh" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.236455 4758 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.320905 4758 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T16:58:07.236487201Z","Handler":null,"Name":""} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.325313 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r5wnr"] Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.326422 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.331394 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.335452 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.335892 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-utilities\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.335979 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-catalog-content\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.336044 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv2xt\" (UniqueName: \"kubernetes.io/projected/d1afe03b-b381-4fc5-973a-f8178760890e-kube-api-access-dv2xt\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: E1203 16:58:07.336187 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:58:07.836161375 +0000 UTC m=+143.037538236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.359202 4758 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.359248 4758 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.410050 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5wnr"] Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.437740 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.437789 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv2xt\" (UniqueName: \"kubernetes.io/projected/d1afe03b-b381-4fc5-973a-f8178760890e-kube-api-access-dv2xt\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.437859 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-utilities\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.437913 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-catalog-content\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.438464 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-catalog-content\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.439379 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-utilities\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.456789 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.456841 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.506856 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv2xt\" (UniqueName: \"kubernetes.io/projected/d1afe03b-b381-4fc5-973a-f8178760890e-kube-api-access-dv2xt\") pod \"redhat-marketplace-r5wnr\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.538792 4758 generic.go:334] "Generic (PLEG): container finished" podID="b43b0669-6da9-438d-bc36-90456cc2611c" containerID="49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf" exitCode=0 Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.539227 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerDied","Data":"49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.539257 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerStarted","Data":"a87273ce688f26abfd57001490d4479b2df9df8458882a8455945015b385b72a"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.547175 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.553885 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94vfh\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.553914 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerStarted","Data":"8fe0ceb326e46ed751427f928aee0097cc359c96d7528f0e2a16d1233935d0f8"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.561901 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l55d4" event={"ID":"8dc530f3-6688-414d-9f8a-bb191ad5a264","Type":"ContainerStarted","Data":"c7dbca67e1c6a55bfdfbb2982734ded077926fbe68883309d6971d34446baff5"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.565254 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.593566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" event={"ID":"c3615454-def8-4637-884b-a4c96fe468d6","Type":"ContainerStarted","Data":"56ea747b3cb7c0c4d09e0f208d8babd53a12e4a60aed672bf697b93e08ae4e22"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.598431 4758 generic.go:334] "Generic (PLEG): container finished" podID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerID="382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86" exitCode=0 Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.599539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerDied","Data":"382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.599566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerStarted","Data":"78da1b48c47d2d2c2619e246a1e57b2e7fcd0edd4f28ac353159315b3b3781a0"} Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.644917 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.719461 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dbcbx"] Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.720782 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.736902 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbcbx"] Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.747641 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcsrz\" (UniqueName: \"kubernetes.io/projected/19fe802e-12eb-45e2-b918-ab7bb933dc04-kube-api-access-hcsrz\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.747868 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-catalog-content\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.747899 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-utilities\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.760788 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.831115 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:07 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:07 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:07 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.831491 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.842885 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.850444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcsrz\" (UniqueName: \"kubernetes.io/projected/19fe802e-12eb-45e2-b918-ab7bb933dc04-kube-api-access-hcsrz\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.850522 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-catalog-content\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.850544 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-utilities\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.851130 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-utilities\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.851771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-catalog-content\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:07 crc kubenswrapper[4758]: I1203 16:58:07.927206 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcsrz\" (UniqueName: \"kubernetes.io/projected/19fe802e-12eb-45e2-b918-ab7bb933dc04-kube-api-access-hcsrz\") pod \"redhat-marketplace-dbcbx\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.046410 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.076953 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.141606 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5wnr"] Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.154911 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbkps\" (UniqueName: \"kubernetes.io/projected/886f030a-c717-4277-b739-393ecb2081e4-kube-api-access-kbkps\") pod \"886f030a-c717-4277-b739-393ecb2081e4\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.155036 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f030a-c717-4277-b739-393ecb2081e4-config-volume\") pod \"886f030a-c717-4277-b739-393ecb2081e4\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.155171 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f030a-c717-4277-b739-393ecb2081e4-secret-volume\") pod \"886f030a-c717-4277-b739-393ecb2081e4\" (UID: \"886f030a-c717-4277-b739-393ecb2081e4\") " Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.156318 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886f030a-c717-4277-b739-393ecb2081e4-config-volume" (OuterVolumeSpecName: "config-volume") pod "886f030a-c717-4277-b739-393ecb2081e4" (UID: "886f030a-c717-4277-b739-393ecb2081e4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.168471 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886f030a-c717-4277-b739-393ecb2081e4-kube-api-access-kbkps" (OuterVolumeSpecName: "kube-api-access-kbkps") pod "886f030a-c717-4277-b739-393ecb2081e4" (UID: "886f030a-c717-4277-b739-393ecb2081e4"). InnerVolumeSpecName "kube-api-access-kbkps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.172725 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/886f030a-c717-4277-b739-393ecb2081e4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "886f030a-c717-4277-b739-393ecb2081e4" (UID: "886f030a-c717-4277-b739-393ecb2081e4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.257292 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f030a-c717-4277-b739-393ecb2081e4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.257713 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f030a-c717-4277-b739-393ecb2081e4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.257723 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbkps\" (UniqueName: \"kubernetes.io/projected/886f030a-c717-4277-b739-393ecb2081e4-kube-api-access-kbkps\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.295098 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94vfh"] Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.428540 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbcbx"] Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.445268 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2xvlb" Dec 03 16:58:08 crc kubenswrapper[4758]: W1203 16:58:08.454712 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19fe802e_12eb_45e2_b918_ab7bb933dc04.slice/crio-77e04033a267a55c995daef9ca6439e34b203cda23f2632891c1b0f9d939c82e WatchSource:0}: Error finding container 77e04033a267a55c995daef9ca6439e34b203cda23f2632891c1b0f9d939c82e: Status 404 returned error can't find the container with id 77e04033a267a55c995daef9ca6439e34b203cda23f2632891c1b0f9d939c82e Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.636777 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" event={"ID":"886f030a-c717-4277-b739-393ecb2081e4","Type":"ContainerDied","Data":"963b01f6bff28045993d91540cfc52ed8041ca3f8e4835d3d714afe976826976"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.637300 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="963b01f6bff28045993d91540cfc52ed8041ca3f8e4835d3d714afe976826976" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.637408 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.673767 4758 generic.go:334] "Generic (PLEG): container finished" podID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerID="f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710" exitCode=0 Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.673871 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerDied","Data":"f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.691335 4758 generic.go:334] "Generic (PLEG): container finished" podID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerID="2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab" exitCode=0 Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.691467 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l55d4" event={"ID":"8dc530f3-6688-414d-9f8a-bb191ad5a264","Type":"ContainerDied","Data":"2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.715938 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1afe03b-b381-4fc5-973a-f8178760890e" containerID="5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa" exitCode=0 Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.716023 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerDied","Data":"5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.716132 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerStarted","Data":"3fc1d4112b366056dc67b50e3efde4ffe77b0f6163d05a5e3946affc76a369f9"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.724632 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8qlms"] Dec 03 16:58:08 crc kubenswrapper[4758]: E1203 16:58:08.725107 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886f030a-c717-4277-b739-393ecb2081e4" containerName="collect-profiles" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.725195 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="886f030a-c717-4277-b739-393ecb2081e4" containerName="collect-profiles" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.725406 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="886f030a-c717-4277-b739-393ecb2081e4" containerName="collect-profiles" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.726572 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.732362 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.746994 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" event={"ID":"c3615454-def8-4637-884b-a4c96fe468d6","Type":"ContainerStarted","Data":"279f76083eb2c93b8a22e778f35184c14f4290ac5ec16e31d8905025e04458da"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.747190 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8qlms"] Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.755058 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbcbx" event={"ID":"19fe802e-12eb-45e2-b918-ab7bb933dc04","Type":"ContainerStarted","Data":"77e04033a267a55c995daef9ca6439e34b203cda23f2632891c1b0f9d939c82e"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.775978 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-utilities\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.776045 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-catalog-content\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.776101 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5pkc\" (UniqueName: \"kubernetes.io/projected/8c7f12da-f734-4617-8977-3004953eb1aa-kube-api-access-j5pkc\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.778776 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" event={"ID":"992baa6d-67aa-467a-8467-c97d64e69d44","Type":"ContainerStarted","Data":"f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.778810 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" event={"ID":"992baa6d-67aa-467a-8467-c97d64e69d44","Type":"ContainerStarted","Data":"2452f2c717c8fc2a8a6b8b83b76e6ea8db99e62dc95a929183175c56b03bd176"} Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.779267 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.784352 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.784411 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.806228 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:08 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:08 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:08 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.806326 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.816608 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.870257 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2gw5x" podStartSLOduration=12.870234373 podStartE2EDuration="12.870234373s" podCreationTimestamp="2025-12-03 16:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:08.869118413 +0000 UTC m=+144.070495294" watchObservedRunningTime="2025-12-03 16:58:08.870234373 +0000 UTC m=+144.071611234" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.877671 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-utilities\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.877771 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-catalog-content\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.877827 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5pkc\" (UniqueName: \"kubernetes.io/projected/8c7f12da-f734-4617-8977-3004953eb1aa-kube-api-access-j5pkc\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.878059 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.878177 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.878782 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-utilities\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.880221 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-catalog-content\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.884494 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.903154 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5pkc\" (UniqueName: \"kubernetes.io/projected/8c7f12da-f734-4617-8977-3004953eb1aa-kube-api-access-j5pkc\") pod \"redhat-operators-8qlms\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.919582 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" podStartSLOduration=125.919557185 podStartE2EDuration="2m5.919557185s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:08.915892418 +0000 UTC m=+144.117269299" watchObservedRunningTime="2025-12-03 16:58:08.919557185 +0000 UTC m=+144.120934046" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.959840 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-qkmml container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.959926 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qkmml" podUID="ceabfcec-18d7-4f41-af6c-b31edfcdbe80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.959860 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-qkmml container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 16:58:08 crc kubenswrapper[4758]: I1203 16:58:08.960002 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qkmml" podUID="ceabfcec-18d7-4f41-af6c-b31edfcdbe80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.053292 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.107991 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vkbzg"] Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.109177 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.129161 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.163513 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkbzg"] Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.285311 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-utilities\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.285848 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-catalog-content\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.285910 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb8tc\" (UniqueName: \"kubernetes.io/projected/adf1ea42-d625-4810-99f9-571f676ae801-kube-api-access-sb8tc\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.348461 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8qlms"] Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.387965 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb8tc\" (UniqueName: \"kubernetes.io/projected/adf1ea42-d625-4810-99f9-571f676ae801-kube-api-access-sb8tc\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.388197 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-utilities\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.388286 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-catalog-content\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.389019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-utilities\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.389059 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-catalog-content\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.415137 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb8tc\" (UniqueName: \"kubernetes.io/projected/adf1ea42-d625-4810-99f9-571f676ae801-kube-api-access-sb8tc\") pod \"redhat-operators-vkbzg\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.428495 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.561264 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.765486 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkbzg"] Dec 03 16:58:09 crc kubenswrapper[4758]: W1203 16:58:09.779711 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadf1ea42_d625_4810_99f9_571f676ae801.slice/crio-5c5085125a3ea6cdc256b9bf5ac52d6df154dfff4e6be5827cd66527b5285d2e WatchSource:0}: Error finding container 5c5085125a3ea6cdc256b9bf5ac52d6df154dfff4e6be5827cd66527b5285d2e: Status 404 returned error can't find the container with id 5c5085125a3ea6cdc256b9bf5ac52d6df154dfff4e6be5827cd66527b5285d2e Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.786977 4758 generic.go:334] "Generic (PLEG): container finished" podID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerID="43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a" exitCode=0 Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.787035 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbcbx" event={"ID":"19fe802e-12eb-45e2-b918-ab7bb933dc04","Type":"ContainerDied","Data":"43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a"} Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.794643 4758 generic.go:334] "Generic (PLEG): container finished" podID="8c7f12da-f734-4617-8977-3004953eb1aa" containerID="2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d" exitCode=0 Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.795894 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerDied","Data":"2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d"} Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.795921 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerStarted","Data":"02b8265c5330f87dbc7e7c5a7e2796160342abba85eeda497088873adbe064a6"} Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.799129 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.801617 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s4h7" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.802413 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-f4wqg" Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.806201 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:09 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:09 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:09 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:09 crc kubenswrapper[4758]: I1203 16:58:09.806278 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.000733 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.001879 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.021601 4758 patch_prober.go:28] interesting pod/console-f9d7485db-kzv6z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.021693 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kzv6z" podUID="a32f32fc-e1aa-402c-93ec-f73446955bd0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.077607 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.078569 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.087342 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.087807 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.088090 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.210863 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d13a84ed-21ac-4a98-bd71-7c45cd041029-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.210940 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d13a84ed-21ac-4a98-bd71-7c45cd041029-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.312687 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d13a84ed-21ac-4a98-bd71-7c45cd041029-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.312781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d13a84ed-21ac-4a98-bd71-7c45cd041029-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.312883 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d13a84ed-21ac-4a98-bd71-7c45cd041029-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.355521 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d13a84ed-21ac-4a98-bd71-7c45cd041029-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.371862 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.372782 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.375266 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.375356 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.377631 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.408742 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.516883 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.516994 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.619150 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.619996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.620158 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.644383 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.699061 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.775494 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 16:58:10 crc kubenswrapper[4758]: W1203 16:58:10.798464 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd13a84ed_21ac_4a98_bd71_7c45cd041029.slice/crio-1693a19cd2cb02ebf47ed28092fa2f39749dfa1c84e90c902e1aceca7f8760bd WatchSource:0}: Error finding container 1693a19cd2cb02ebf47ed28092fa2f39749dfa1c84e90c902e1aceca7f8760bd: Status 404 returned error can't find the container with id 1693a19cd2cb02ebf47ed28092fa2f39749dfa1c84e90c902e1aceca7f8760bd Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.806718 4758 generic.go:334] "Generic (PLEG): container finished" podID="adf1ea42-d625-4810-99f9-571f676ae801" containerID="62a1166edcf9331d311ee271b598a8c1fa9b6ac5d3bc0df7c7adde3957c9f80a" exitCode=0 Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.806918 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerDied","Data":"62a1166edcf9331d311ee271b598a8c1fa9b6ac5d3bc0df7c7adde3957c9f80a"} Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.807006 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerStarted","Data":"5c5085125a3ea6cdc256b9bf5ac52d6df154dfff4e6be5827cd66527b5285d2e"} Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.815898 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:10 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:10 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:10 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:10 crc kubenswrapper[4758]: I1203 16:58:10.815957 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.026843 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.027307 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.034093 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.128853 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.128936 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.129670 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.136660 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.138427 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.151936 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 16:58:11 crc kubenswrapper[4758]: W1203 16:58:11.184223 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda6373ce8_3da7_4568_9aa6_f02c676e8c4e.slice/crio-c6b9588e42ab2b9769b1d6a0ac71cdeabee2569b81522d41d052e6885d6f9826 WatchSource:0}: Error finding container c6b9588e42ab2b9769b1d6a0ac71cdeabee2569b81522d41d052e6885d6f9826: Status 404 returned error can't find the container with id c6b9588e42ab2b9769b1d6a0ac71cdeabee2569b81522d41d052e6885d6f9826 Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.330259 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.337776 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.345065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.395314 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.395858 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:58:11 crc kubenswrapper[4758]: W1203 16:58:11.769576 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-d4d4fdb4aa72d6903ac2622fde6f9e80033eb0a986f727ad37c02b3e801e195d WatchSource:0}: Error finding container d4d4fdb4aa72d6903ac2622fde6f9e80033eb0a986f727ad37c02b3e801e195d: Status 404 returned error can't find the container with id d4d4fdb4aa72d6903ac2622fde6f9e80033eb0a986f727ad37c02b3e801e195d Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.805484 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:11 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Dec 03 16:58:11 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:11 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.806078 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.845941 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d13a84ed-21ac-4a98-bd71-7c45cd041029","Type":"ContainerStarted","Data":"1693a19cd2cb02ebf47ed28092fa2f39749dfa1c84e90c902e1aceca7f8760bd"} Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.890540 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6373ce8-3da7-4568-9aa6-f02c676e8c4e","Type":"ContainerStarted","Data":"c6b9588e42ab2b9769b1d6a0ac71cdeabee2569b81522d41d052e6885d6f9826"} Dec 03 16:58:11 crc kubenswrapper[4758]: I1203 16:58:11.907459 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d4d4fdb4aa72d6903ac2622fde6f9e80033eb0a986f727ad37c02b3e801e195d"} Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.804920 4758 patch_prober.go:28] interesting pod/router-default-5444994796-jxqhr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:58:12 crc kubenswrapper[4758]: [+]has-synced ok Dec 03 16:58:12 crc kubenswrapper[4758]: [+]process-running ok Dec 03 16:58:12 crc kubenswrapper[4758]: healthz check failed Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.805478 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxqhr" podUID="58900711-350a-4753-888d-7a6ff520abd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.942931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a7e76c3662b33ed72a3458ae864b22a866ca17c7d3d96f70979e389e5d2a6c9b"} Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.947168 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5aaff5966f506a490c320c2e68aec1f14c900a2b7468c1a5ac2b33f8dda05b79"} Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.949507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9a869e6e4c33cb7d0bce183c6f3d496e58e315fe6b0830c88ccc4900834df7f8"} Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.955927 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d13a84ed-21ac-4a98-bd71-7c45cd041029","Type":"ContainerStarted","Data":"6aed1cae65409dfa7c90bcf74b38f40ccb51dde4da3fe91b358d408be9c7383c"} Dec 03 16:58:12 crc kubenswrapper[4758]: I1203 16:58:12.962715 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6373ce8-3da7-4568-9aa6-f02c676e8c4e","Type":"ContainerStarted","Data":"d23abade4f5d774083a1a7262be26f8fcd67950a6cb6e2eec5d2d5fd1730487a"} Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.804998 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.818383 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jxqhr" Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.985996 4758 generic.go:334] "Generic (PLEG): container finished" podID="a6373ce8-3da7-4568-9aa6-f02c676e8c4e" containerID="d23abade4f5d774083a1a7262be26f8fcd67950a6cb6e2eec5d2d5fd1730487a" exitCode=0 Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.986106 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6373ce8-3da7-4568-9aa6-f02c676e8c4e","Type":"ContainerDied","Data":"d23abade4f5d774083a1a7262be26f8fcd67950a6cb6e2eec5d2d5fd1730487a"} Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.990068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"01dea652151413bc1b77495ac6e0c5e6d9c591bfc0bbdb466b54937732ca109a"} Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.995872 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6d2d0fc2b965f41d3657eae5b3748bd34f404eba5a94ac056a0fe547cdae0101"} Dec 03 16:58:13 crc kubenswrapper[4758]: I1203 16:58:13.996047 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:13.999525 4758 generic.go:334] "Generic (PLEG): container finished" podID="d13a84ed-21ac-4a98-bd71-7c45cd041029" containerID="6aed1cae65409dfa7c90bcf74b38f40ccb51dde4da3fe91b358d408be9c7383c" exitCode=0 Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:13.999662 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d13a84ed-21ac-4a98-bd71-7c45cd041029","Type":"ContainerDied","Data":"6aed1cae65409dfa7c90bcf74b38f40ccb51dde4da3fe91b358d408be9c7383c"} Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.382906 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.391437 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.518512 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d13a84ed-21ac-4a98-bd71-7c45cd041029-kube-api-access\") pod \"d13a84ed-21ac-4a98-bd71-7c45cd041029\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.518568 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d13a84ed-21ac-4a98-bd71-7c45cd041029-kubelet-dir\") pod \"d13a84ed-21ac-4a98-bd71-7c45cd041029\" (UID: \"d13a84ed-21ac-4a98-bd71-7c45cd041029\") " Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.518650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kubelet-dir\") pod \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.518753 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kube-api-access\") pod \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\" (UID: \"a6373ce8-3da7-4568-9aa6-f02c676e8c4e\") " Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.521095 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d13a84ed-21ac-4a98-bd71-7c45cd041029-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d13a84ed-21ac-4a98-bd71-7c45cd041029" (UID: "d13a84ed-21ac-4a98-bd71-7c45cd041029"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.521186 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a6373ce8-3da7-4568-9aa6-f02c676e8c4e" (UID: "a6373ce8-3da7-4568-9aa6-f02c676e8c4e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.528006 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a6373ce8-3da7-4568-9aa6-f02c676e8c4e" (UID: "a6373ce8-3da7-4568-9aa6-f02c676e8c4e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.528480 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d13a84ed-21ac-4a98-bd71-7c45cd041029-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d13a84ed-21ac-4a98-bd71-7c45cd041029" (UID: "d13a84ed-21ac-4a98-bd71-7c45cd041029"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.621612 4758 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d13a84ed-21ac-4a98-bd71-7c45cd041029-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.621657 4758 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.621668 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6373ce8-3da7-4568-9aa6-f02c676e8c4e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.621696 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d13a84ed-21ac-4a98-bd71-7c45cd041029-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:14 crc kubenswrapper[4758]: I1203 16:58:14.876112 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-c9b57" Dec 03 16:58:15 crc kubenswrapper[4758]: I1203 16:58:15.030158 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:58:15 crc kubenswrapper[4758]: I1203 16:58:15.035829 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d13a84ed-21ac-4a98-bd71-7c45cd041029","Type":"ContainerDied","Data":"1693a19cd2cb02ebf47ed28092fa2f39749dfa1c84e90c902e1aceca7f8760bd"} Dec 03 16:58:15 crc kubenswrapper[4758]: I1203 16:58:15.035877 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1693a19cd2cb02ebf47ed28092fa2f39749dfa1c84e90c902e1aceca7f8760bd" Dec 03 16:58:15 crc kubenswrapper[4758]: I1203 16:58:15.044328 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6373ce8-3da7-4568-9aa6-f02c676e8c4e","Type":"ContainerDied","Data":"c6b9588e42ab2b9769b1d6a0ac71cdeabee2569b81522d41d052e6885d6f9826"} Dec 03 16:58:15 crc kubenswrapper[4758]: I1203 16:58:15.044402 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6b9588e42ab2b9769b1d6a0ac71cdeabee2569b81522d41d052e6885d6f9826" Dec 03 16:58:15 crc kubenswrapper[4758]: I1203 16:58:15.044704 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:58:18 crc kubenswrapper[4758]: I1203 16:58:18.960901 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qkmml" Dec 03 16:58:20 crc kubenswrapper[4758]: I1203 16:58:20.096299 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:58:20 crc kubenswrapper[4758]: I1203 16:58:20.099991 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 16:58:26 crc kubenswrapper[4758]: I1203 16:58:26.356268 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:58:26 crc kubenswrapper[4758]: I1203 16:58:26.368661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af-metrics-certs\") pod \"network-metrics-daemon-55dnk\" (UID: \"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af\") " pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:58:26 crc kubenswrapper[4758]: I1203 16:58:26.651565 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-55dnk" Dec 03 16:58:27 crc kubenswrapper[4758]: I1203 16:58:27.850613 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 16:58:39 crc kubenswrapper[4758]: E1203 16:58:39.202307 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 16:58:39 crc kubenswrapper[4758]: E1203 16:58:39.203369 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j5pkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-8qlms_openshift-marketplace(8c7f12da-f734-4617-8977-3004953eb1aa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:39 crc kubenswrapper[4758]: E1203 16:58:39.204558 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-8qlms" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" Dec 03 16:58:39 crc kubenswrapper[4758]: I1203 16:58:39.826448 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xsgd2" Dec 03 16:58:41 crc kubenswrapper[4758]: I1203 16:58:41.394835 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:58:41 crc kubenswrapper[4758]: I1203 16:58:41.395287 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.683077 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 16:58:44 crc kubenswrapper[4758]: E1203 16:58:44.684935 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d13a84ed-21ac-4a98-bd71-7c45cd041029" containerName="pruner" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.685144 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d13a84ed-21ac-4a98-bd71-7c45cd041029" containerName="pruner" Dec 03 16:58:44 crc kubenswrapper[4758]: E1203 16:58:44.685350 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6373ce8-3da7-4568-9aa6-f02c676e8c4e" containerName="pruner" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.686841 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6373ce8-3da7-4568-9aa6-f02c676e8c4e" containerName="pruner" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.687061 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d13a84ed-21ac-4a98-bd71-7c45cd041029" containerName="pruner" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.687083 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6373ce8-3da7-4568-9aa6-f02c676e8c4e" containerName="pruner" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.687700 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.687846 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.691260 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.693121 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.740144 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.740216 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.842060 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.841953 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.842225 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:44 crc kubenswrapper[4758]: I1203 16:58:44.997417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:45 crc kubenswrapper[4758]: I1203 16:58:45.290150 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 16:58:45 crc kubenswrapper[4758]: I1203 16:58:45.297634 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:46 crc kubenswrapper[4758]: E1203 16:58:46.661119 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-8qlms" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" Dec 03 16:58:47 crc kubenswrapper[4758]: E1203 16:58:47.089269 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 16:58:47 crc kubenswrapper[4758]: E1203 16:58:47.089935 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ms74c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d4b9m_openshift-marketplace(b43b0669-6da9-438d-bc36-90456cc2611c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:47 crc kubenswrapper[4758]: E1203 16:58:47.091184 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-d4b9m" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" Dec 03 16:58:47 crc kubenswrapper[4758]: E1203 16:58:47.204737 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 16:58:47 crc kubenswrapper[4758]: E1203 16:58:47.204935 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-92f86,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jm64k_openshift-marketplace(db60029f-ec45-4c51-98cd-ca184c65c38d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:47 crc kubenswrapper[4758]: E1203 16:58:47.206093 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jm64k" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" Dec 03 16:58:48 crc kubenswrapper[4758]: E1203 16:58:48.243396 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d4b9m" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" Dec 03 16:58:48 crc kubenswrapper[4758]: E1203 16:58:48.244192 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jm64k" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" Dec 03 16:58:48 crc kubenswrapper[4758]: E1203 16:58:48.302555 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 16:58:48 crc kubenswrapper[4758]: E1203 16:58:48.302791 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hcsrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dbcbx_openshift-marketplace(19fe802e-12eb-45e2-b918-ab7bb933dc04): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:48 crc kubenswrapper[4758]: E1203 16:58:48.303968 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dbcbx" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.280126 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.281237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.299569 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.304714 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e7787c1-2384-45d9-a620-096dd4ed60e0-kube-api-access\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.304769 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-var-lock\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.304872 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.406086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.406179 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e7787c1-2384-45d9-a620-096dd4ed60e0-kube-api-access\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.406214 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-var-lock\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.406345 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-var-lock\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.406403 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.432819 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e7787c1-2384-45d9-a620-096dd4ed60e0-kube-api-access\") pod \"installer-9-crc\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: I1203 16:58:49.611123 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.755874 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dbcbx" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.849368 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.849581 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dv2xt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-r5wnr_openshift-marketplace(d1afe03b-b381-4fc5-973a-f8178760890e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.851027 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-r5wnr" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.852104 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.852316 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dnbjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7kgk7_openshift-marketplace(f673f32d-f81f-4fe9-ab86-801fe50a46da): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.853752 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7kgk7" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.942098 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.942800 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thv2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-l55d4_openshift-marketplace(8dc530f3-6688-414d-9f8a-bb191ad5a264): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:58:49 crc kubenswrapper[4758]: E1203 16:58:49.944377 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-l55d4" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.071232 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.147958 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-55dnk"] Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.247956 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.325011 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerStarted","Data":"effa7e3d486f8c4c8c329202c6de90f1c22236531edc809adba3fa7fc896743d"} Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.326817 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-55dnk" event={"ID":"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af","Type":"ContainerStarted","Data":"04d719dc86da00ea888a1e2de5f313e79e8abd77a51fab39e085067f4565d8b0"} Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.328400 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0e7787c1-2384-45d9-a620-096dd4ed60e0","Type":"ContainerStarted","Data":"d69e27ca93ec2a6c1020af8c2a29d4821eff3f82719baf5aaa0092a669034157"} Dec 03 16:58:50 crc kubenswrapper[4758]: I1203 16:58:50.333318 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4fdc199-9fb2-48bb-81a7-aab2454867eb","Type":"ContainerStarted","Data":"37802d58e1dcc71570031bff483357f6ef14ece223630dd10c850b4f63047461"} Dec 03 16:58:50 crc kubenswrapper[4758]: E1203 16:58:50.335672 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-r5wnr" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" Dec 03 16:58:50 crc kubenswrapper[4758]: E1203 16:58:50.336206 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-l55d4" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" Dec 03 16:58:50 crc kubenswrapper[4758]: E1203 16:58:50.336547 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7kgk7" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.341540 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.346128 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4fdc199-9fb2-48bb-81a7-aab2454867eb","Type":"ContainerStarted","Data":"372df3cfef7a2b5178b0aeb408f6078ed1f3fcebdbc45ab0224e4b4df0dbc663"} Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.351507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0e7787c1-2384-45d9-a620-096dd4ed60e0","Type":"ContainerStarted","Data":"dc23826b9bf8bc185f2a8a0189b5c11746f7f048e0d4ed78eba5925e711c954a"} Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.355175 4758 generic.go:334] "Generic (PLEG): container finished" podID="adf1ea42-d625-4810-99f9-571f676ae801" containerID="effa7e3d486f8c4c8c329202c6de90f1c22236531edc809adba3fa7fc896743d" exitCode=0 Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.355259 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerDied","Data":"effa7e3d486f8c4c8c329202c6de90f1c22236531edc809adba3fa7fc896743d"} Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.359354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-55dnk" event={"ID":"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af","Type":"ContainerStarted","Data":"534d903573cc4eee104dfd3f0185ae96176d70682cd9230bc8df7d399de1dfc2"} Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.359402 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-55dnk" event={"ID":"5fcc3acd-2e30-49d7-bab1-4a10c1f2a9af","Type":"ContainerStarted","Data":"d7d3e27451bfd10b65053062fabdd8a71ae00b2f5c2c249c60350d9e1c29e4b2"} Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.404918 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-55dnk" podStartSLOduration=168.404893004 podStartE2EDuration="2m48.404893004s" podCreationTimestamp="2025-12-03 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:51.401830302 +0000 UTC m=+186.603207163" watchObservedRunningTime="2025-12-03 16:58:51.404893004 +0000 UTC m=+186.606269875" Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.426100 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.426071862 podStartE2EDuration="2.426071862s" podCreationTimestamp="2025-12-03 16:58:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:51.42079643 +0000 UTC m=+186.622173301" watchObservedRunningTime="2025-12-03 16:58:51.426071862 +0000 UTC m=+186.627448723" Dec 03 16:58:51 crc kubenswrapper[4758]: I1203 16:58:51.444706 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=7.44466411 podStartE2EDuration="7.44466411s" podCreationTimestamp="2025-12-03 16:58:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:58:51.441413233 +0000 UTC m=+186.642790094" watchObservedRunningTime="2025-12-03 16:58:51.44466411 +0000 UTC m=+186.646040981" Dec 03 16:58:52 crc kubenswrapper[4758]: I1203 16:58:52.366039 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4fdc199-9fb2-48bb-81a7-aab2454867eb" containerID="372df3cfef7a2b5178b0aeb408f6078ed1f3fcebdbc45ab0224e4b4df0dbc663" exitCode=0 Dec 03 16:58:52 crc kubenswrapper[4758]: I1203 16:58:52.367033 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4fdc199-9fb2-48bb-81a7-aab2454867eb","Type":"ContainerDied","Data":"372df3cfef7a2b5178b0aeb408f6078ed1f3fcebdbc45ab0224e4b4df0dbc663"} Dec 03 16:58:52 crc kubenswrapper[4758]: I1203 16:58:52.371100 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerStarted","Data":"3a5c481956f4890d2354770e1c83a76ac07fec04f73dc9bae23a9b2f1f986d6b"} Dec 03 16:58:52 crc kubenswrapper[4758]: I1203 16:58:52.412537 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vkbzg" podStartSLOduration=2.479688331 podStartE2EDuration="43.412511905s" podCreationTimestamp="2025-12-03 16:58:09 +0000 UTC" firstStartedPulling="2025-12-03 16:58:10.832520505 +0000 UTC m=+146.033897366" lastFinishedPulling="2025-12-03 16:58:51.765344089 +0000 UTC m=+186.966720940" observedRunningTime="2025-12-03 16:58:52.406716569 +0000 UTC m=+187.608093430" watchObservedRunningTime="2025-12-03 16:58:52.412511905 +0000 UTC m=+187.613888766" Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.669474 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.681007 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kubelet-dir\") pod \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.681136 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kube-api-access\") pod \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\" (UID: \"a4fdc199-9fb2-48bb-81a7-aab2454867eb\") " Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.681813 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a4fdc199-9fb2-48bb-81a7-aab2454867eb" (UID: "a4fdc199-9fb2-48bb-81a7-aab2454867eb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.696446 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a4fdc199-9fb2-48bb-81a7-aab2454867eb" (UID: "a4fdc199-9fb2-48bb-81a7-aab2454867eb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.783352 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:53 crc kubenswrapper[4758]: I1203 16:58:53.783515 4758 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4fdc199-9fb2-48bb-81a7-aab2454867eb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:54 crc kubenswrapper[4758]: I1203 16:58:54.384244 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4fdc199-9fb2-48bb-81a7-aab2454867eb","Type":"ContainerDied","Data":"37802d58e1dcc71570031bff483357f6ef14ece223630dd10c850b4f63047461"} Dec 03 16:58:54 crc kubenswrapper[4758]: I1203 16:58:54.384307 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:58:54 crc kubenswrapper[4758]: I1203 16:58:54.384318 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37802d58e1dcc71570031bff483357f6ef14ece223630dd10c850b4f63047461" Dec 03 16:58:58 crc kubenswrapper[4758]: I1203 16:58:58.118052 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nqrwv"] Dec 03 16:58:59 crc kubenswrapper[4758]: I1203 16:58:59.429505 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:59 crc kubenswrapper[4758]: I1203 16:58:59.430065 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:58:59 crc kubenswrapper[4758]: I1203 16:58:59.933234 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:59:00 crc kubenswrapper[4758]: I1203 16:59:00.422223 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerStarted","Data":"6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e"} Dec 03 16:59:00 crc kubenswrapper[4758]: I1203 16:59:00.481868 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:59:01 crc kubenswrapper[4758]: I1203 16:59:01.172547 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkbzg"] Dec 03 16:59:02 crc kubenswrapper[4758]: I1203 16:59:02.435110 4758 generic.go:334] "Generic (PLEG): container finished" podID="8c7f12da-f734-4617-8977-3004953eb1aa" containerID="6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e" exitCode=0 Dec 03 16:59:02 crc kubenswrapper[4758]: I1203 16:59:02.435221 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerDied","Data":"6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e"} Dec 03 16:59:02 crc kubenswrapper[4758]: I1203 16:59:02.435816 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vkbzg" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="registry-server" containerID="cri-o://3a5c481956f4890d2354770e1c83a76ac07fec04f73dc9bae23a9b2f1f986d6b" gracePeriod=2 Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.449202 4758 generic.go:334] "Generic (PLEG): container finished" podID="adf1ea42-d625-4810-99f9-571f676ae801" containerID="3a5c481956f4890d2354770e1c83a76ac07fec04f73dc9bae23a9b2f1f986d6b" exitCode=0 Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.449280 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerDied","Data":"3a5c481956f4890d2354770e1c83a76ac07fec04f73dc9bae23a9b2f1f986d6b"} Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.865418 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.916392 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-catalog-content\") pod \"adf1ea42-d625-4810-99f9-571f676ae801\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.916468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-utilities\") pod \"adf1ea42-d625-4810-99f9-571f676ae801\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.916567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb8tc\" (UniqueName: \"kubernetes.io/projected/adf1ea42-d625-4810-99f9-571f676ae801-kube-api-access-sb8tc\") pod \"adf1ea42-d625-4810-99f9-571f676ae801\" (UID: \"adf1ea42-d625-4810-99f9-571f676ae801\") " Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.917790 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-utilities" (OuterVolumeSpecName: "utilities") pod "adf1ea42-d625-4810-99f9-571f676ae801" (UID: "adf1ea42-d625-4810-99f9-571f676ae801"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:04 crc kubenswrapper[4758]: I1203 16:59:04.925977 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf1ea42-d625-4810-99f9-571f676ae801-kube-api-access-sb8tc" (OuterVolumeSpecName: "kube-api-access-sb8tc") pod "adf1ea42-d625-4810-99f9-571f676ae801" (UID: "adf1ea42-d625-4810-99f9-571f676ae801"). InnerVolumeSpecName "kube-api-access-sb8tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.018575 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb8tc\" (UniqueName: \"kubernetes.io/projected/adf1ea42-d625-4810-99f9-571f676ae801-kube-api-access-sb8tc\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.018613 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.067805 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adf1ea42-d625-4810-99f9-571f676ae801" (UID: "adf1ea42-d625-4810-99f9-571f676ae801"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.119979 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf1ea42-d625-4810-99f9-571f676ae801-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.459350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerStarted","Data":"fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.468590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerStarted","Data":"83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.471075 4758 generic.go:334] "Generic (PLEG): container finished" podID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerID="b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8" exitCode=0 Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.471124 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l55d4" event={"ID":"8dc530f3-6688-414d-9f8a-bb191ad5a264","Type":"ContainerDied","Data":"b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.475652 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerStarted","Data":"fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.488447 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkbzg" event={"ID":"adf1ea42-d625-4810-99f9-571f676ae801","Type":"ContainerDied","Data":"5c5085125a3ea6cdc256b9bf5ac52d6df154dfff4e6be5827cd66527b5285d2e"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.488544 4758 scope.go:117] "RemoveContainer" containerID="3a5c481956f4890d2354770e1c83a76ac07fec04f73dc9bae23a9b2f1f986d6b" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.488860 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkbzg" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.494533 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerStarted","Data":"18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.513963 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerStarted","Data":"2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf"} Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.515375 4758 scope.go:117] "RemoveContainer" containerID="effa7e3d486f8c4c8c329202c6de90f1c22236531edc809adba3fa7fc896743d" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.535196 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8qlms" podStartSLOduration=2.568106214 podStartE2EDuration="57.535177673s" podCreationTimestamp="2025-12-03 16:58:08 +0000 UTC" firstStartedPulling="2025-12-03 16:58:09.815327537 +0000 UTC m=+145.016704398" lastFinishedPulling="2025-12-03 16:59:04.782398996 +0000 UTC m=+199.983775857" observedRunningTime="2025-12-03 16:59:05.501640634 +0000 UTC m=+200.703017495" watchObservedRunningTime="2025-12-03 16:59:05.535177673 +0000 UTC m=+200.736554534" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.554101 4758 scope.go:117] "RemoveContainer" containerID="62a1166edcf9331d311ee271b598a8c1fa9b6ac5d3bc0df7c7adde3957c9f80a" Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.655251 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkbzg"] Dec 03 16:59:05 crc kubenswrapper[4758]: I1203 16:59:05.658358 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vkbzg"] Dec 03 16:59:06 crc kubenswrapper[4758]: I1203 16:59:06.523290 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1afe03b-b381-4fc5-973a-f8178760890e" containerID="fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43" exitCode=0 Dec 03 16:59:06 crc kubenswrapper[4758]: I1203 16:59:06.523808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerDied","Data":"fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43"} Dec 03 16:59:06 crc kubenswrapper[4758]: I1203 16:59:06.529697 4758 generic.go:334] "Generic (PLEG): container finished" podID="b43b0669-6da9-438d-bc36-90456cc2611c" containerID="2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf" exitCode=0 Dec 03 16:59:06 crc kubenswrapper[4758]: I1203 16:59:06.529731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerDied","Data":"2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.125198 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf1ea42-d625-4810-99f9-571f676ae801" path="/var/lib/kubelet/pods/adf1ea42-d625-4810-99f9-571f676ae801/volumes" Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.538172 4758 generic.go:334] "Generic (PLEG): container finished" podID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerID="18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b" exitCode=0 Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.538241 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerDied","Data":"18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.541632 4758 generic.go:334] "Generic (PLEG): container finished" podID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerID="9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80" exitCode=0 Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.541721 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbcbx" event={"ID":"19fe802e-12eb-45e2-b918-ab7bb933dc04","Type":"ContainerDied","Data":"9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.544297 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerStarted","Data":"72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.574768 4758 generic.go:334] "Generic (PLEG): container finished" podID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerID="83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58" exitCode=0 Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.575188 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerDied","Data":"83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.590706 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l55d4" event={"ID":"8dc530f3-6688-414d-9f8a-bb191ad5a264","Type":"ContainerStarted","Data":"09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.593956 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerStarted","Data":"7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d"} Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.618412 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d4b9m" podStartSLOduration=3.119317494 podStartE2EDuration="1m2.618388476s" podCreationTimestamp="2025-12-03 16:58:05 +0000 UTC" firstStartedPulling="2025-12-03 16:58:07.546899546 +0000 UTC m=+142.748276407" lastFinishedPulling="2025-12-03 16:59:07.045970528 +0000 UTC m=+202.247347389" observedRunningTime="2025-12-03 16:59:07.617141742 +0000 UTC m=+202.818518603" watchObservedRunningTime="2025-12-03 16:59:07.618388476 +0000 UTC m=+202.819765337" Dec 03 16:59:07 crc kubenswrapper[4758]: I1203 16:59:07.684954 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r5wnr" podStartSLOduration=2.299201454 podStartE2EDuration="1m0.684918109s" podCreationTimestamp="2025-12-03 16:58:07 +0000 UTC" firstStartedPulling="2025-12-03 16:58:08.724878525 +0000 UTC m=+143.926255396" lastFinishedPulling="2025-12-03 16:59:07.1105952 +0000 UTC m=+202.311972051" observedRunningTime="2025-12-03 16:59:07.681169978 +0000 UTC m=+202.882546839" watchObservedRunningTime="2025-12-03 16:59:07.684918109 +0000 UTC m=+202.886294970" Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.601636 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerStarted","Data":"cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518"} Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.606104 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbcbx" event={"ID":"19fe802e-12eb-45e2-b918-ab7bb933dc04","Type":"ContainerStarted","Data":"f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081"} Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.609791 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerStarted","Data":"3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e"} Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.637440 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7kgk7" podStartSLOduration=3.331058165 podStartE2EDuration="1m3.63741692s" podCreationTimestamp="2025-12-03 16:58:05 +0000 UTC" firstStartedPulling="2025-12-03 16:58:07.609014461 +0000 UTC m=+142.810391322" lastFinishedPulling="2025-12-03 16:59:07.915373216 +0000 UTC m=+203.116750077" observedRunningTime="2025-12-03 16:59:08.63007962 +0000 UTC m=+203.831456481" watchObservedRunningTime="2025-12-03 16:59:08.63741692 +0000 UTC m=+203.838793781" Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.639480 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l55d4" podStartSLOduration=5.570804272 podStartE2EDuration="1m3.639470815s" podCreationTimestamp="2025-12-03 16:58:05 +0000 UTC" firstStartedPulling="2025-12-03 16:58:08.701888369 +0000 UTC m=+143.903265220" lastFinishedPulling="2025-12-03 16:59:06.770554882 +0000 UTC m=+201.971931763" observedRunningTime="2025-12-03 16:59:07.707248334 +0000 UTC m=+202.908625195" watchObservedRunningTime="2025-12-03 16:59:08.639470815 +0000 UTC m=+203.840847676" Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.654671 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jm64k" podStartSLOduration=4.299071572 podStartE2EDuration="1m3.654648477s" podCreationTimestamp="2025-12-03 16:58:05 +0000 UTC" firstStartedPulling="2025-12-03 16:58:08.685321834 +0000 UTC m=+143.886698695" lastFinishedPulling="2025-12-03 16:59:08.040898749 +0000 UTC m=+203.242275600" observedRunningTime="2025-12-03 16:59:08.653437374 +0000 UTC m=+203.854814235" watchObservedRunningTime="2025-12-03 16:59:08.654648477 +0000 UTC m=+203.856025338" Dec 03 16:59:08 crc kubenswrapper[4758]: I1203 16:59:08.683164 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dbcbx" podStartSLOduration=3.4441791950000002 podStartE2EDuration="1m1.683134829s" podCreationTimestamp="2025-12-03 16:58:07 +0000 UTC" firstStartedPulling="2025-12-03 16:58:09.815463531 +0000 UTC m=+145.016840402" lastFinishedPulling="2025-12-03 16:59:08.054419175 +0000 UTC m=+203.255796036" observedRunningTime="2025-12-03 16:59:08.678946865 +0000 UTC m=+203.880323726" watchObservedRunningTime="2025-12-03 16:59:08.683134829 +0000 UTC m=+203.884511690" Dec 03 16:59:09 crc kubenswrapper[4758]: I1203 16:59:09.054945 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:59:09 crc kubenswrapper[4758]: I1203 16:59:09.055019 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:59:10 crc kubenswrapper[4758]: I1203 16:59:10.103628 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8qlms" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="registry-server" probeResult="failure" output=< Dec 03 16:59:10 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 16:59:10 crc kubenswrapper[4758]: > Dec 03 16:59:11 crc kubenswrapper[4758]: I1203 16:59:11.394562 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:59:11 crc kubenswrapper[4758]: I1203 16:59:11.394649 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:59:11 crc kubenswrapper[4758]: I1203 16:59:11.394731 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 16:59:11 crc kubenswrapper[4758]: I1203 16:59:11.395437 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:59:11 crc kubenswrapper[4758]: I1203 16:59:11.395551 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259" gracePeriod=600 Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.663738 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259" exitCode=0 Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.663812 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259"} Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.687251 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.687320 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.732527 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.870563 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.870625 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:59:15 crc kubenswrapper[4758]: I1203 16:59:15.919848 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.281663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.282178 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.326935 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.403882 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.403958 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.476259 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.718571 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.719522 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.747411 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 16:59:16 crc kubenswrapper[4758]: I1203 16:59:16.897996 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 16:59:17 crc kubenswrapper[4758]: I1203 16:59:17.368964 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jm64k"] Dec 03 16:59:17 crc kubenswrapper[4758]: I1203 16:59:17.566152 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:59:17 crc kubenswrapper[4758]: I1203 16:59:17.566232 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:59:17 crc kubenswrapper[4758]: I1203 16:59:17.607676 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:59:17 crc kubenswrapper[4758]: I1203 16:59:17.678531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"64765cfa70b0509666919f46ee9d2002e45bb95cfbe37b712a0e15a96f9a4223"} Dec 03 16:59:17 crc kubenswrapper[4758]: I1203 16:59:17.748383 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.047371 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.048057 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.119104 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.686119 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jm64k" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="registry-server" containerID="cri-o://3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e" gracePeriod=2 Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.774998 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l55d4"] Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.775389 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l55d4" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="registry-server" containerID="cri-o://09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268" gracePeriod=2 Dec 03 16:59:18 crc kubenswrapper[4758]: I1203 16:59:18.802981 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.103636 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.146233 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.564040 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.652143 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.656941 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-utilities\") pod \"db60029f-ec45-4c51-98cd-ca184c65c38d\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.657006 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-catalog-content\") pod \"db60029f-ec45-4c51-98cd-ca184c65c38d\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.657099 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92f86\" (UniqueName: \"kubernetes.io/projected/db60029f-ec45-4c51-98cd-ca184c65c38d-kube-api-access-92f86\") pod \"db60029f-ec45-4c51-98cd-ca184c65c38d\" (UID: \"db60029f-ec45-4c51-98cd-ca184c65c38d\") " Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.657129 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-catalog-content\") pod \"8dc530f3-6688-414d-9f8a-bb191ad5a264\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.657204 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thv2f\" (UniqueName: \"kubernetes.io/projected/8dc530f3-6688-414d-9f8a-bb191ad5a264-kube-api-access-thv2f\") pod \"8dc530f3-6688-414d-9f8a-bb191ad5a264\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.657232 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-utilities\") pod \"8dc530f3-6688-414d-9f8a-bb191ad5a264\" (UID: \"8dc530f3-6688-414d-9f8a-bb191ad5a264\") " Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.657694 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-utilities" (OuterVolumeSpecName: "utilities") pod "db60029f-ec45-4c51-98cd-ca184c65c38d" (UID: "db60029f-ec45-4c51-98cd-ca184c65c38d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.658185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-utilities" (OuterVolumeSpecName: "utilities") pod "8dc530f3-6688-414d-9f8a-bb191ad5a264" (UID: "8dc530f3-6688-414d-9f8a-bb191ad5a264"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.664271 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc530f3-6688-414d-9f8a-bb191ad5a264-kube-api-access-thv2f" (OuterVolumeSpecName: "kube-api-access-thv2f") pod "8dc530f3-6688-414d-9f8a-bb191ad5a264" (UID: "8dc530f3-6688-414d-9f8a-bb191ad5a264"). InnerVolumeSpecName "kube-api-access-thv2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.687213 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db60029f-ec45-4c51-98cd-ca184c65c38d-kube-api-access-92f86" (OuterVolumeSpecName: "kube-api-access-92f86") pod "db60029f-ec45-4c51-98cd-ca184c65c38d" (UID: "db60029f-ec45-4c51-98cd-ca184c65c38d"). InnerVolumeSpecName "kube-api-access-92f86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.706036 4758 generic.go:334] "Generic (PLEG): container finished" podID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerID="3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e" exitCode=0 Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.706217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerDied","Data":"3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e"} Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.706251 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm64k" event={"ID":"db60029f-ec45-4c51-98cd-ca184c65c38d","Type":"ContainerDied","Data":"8fe0ceb326e46ed751427f928aee0097cc359c96d7528f0e2a16d1233935d0f8"} Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.706271 4758 scope.go:117] "RemoveContainer" containerID="3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.706409 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm64k" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.712775 4758 generic.go:334] "Generic (PLEG): container finished" podID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerID="09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268" exitCode=0 Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.713560 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l55d4" event={"ID":"8dc530f3-6688-414d-9f8a-bb191ad5a264","Type":"ContainerDied","Data":"09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268"} Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.713609 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l55d4" event={"ID":"8dc530f3-6688-414d-9f8a-bb191ad5a264","Type":"ContainerDied","Data":"c7dbca67e1c6a55bfdfbb2982734ded077926fbe68883309d6971d34446baff5"} Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.713704 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l55d4" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.736288 4758 scope.go:117] "RemoveContainer" containerID="83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.757921 4758 scope.go:117] "RemoveContainer" containerID="f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.758617 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92f86\" (UniqueName: \"kubernetes.io/projected/db60029f-ec45-4c51-98cd-ca184c65c38d-kube-api-access-92f86\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.758639 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thv2f\" (UniqueName: \"kubernetes.io/projected/8dc530f3-6688-414d-9f8a-bb191ad5a264-kube-api-access-thv2f\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.758651 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.758663 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.766743 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dc530f3-6688-414d-9f8a-bb191ad5a264" (UID: "8dc530f3-6688-414d-9f8a-bb191ad5a264"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.782860 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db60029f-ec45-4c51-98cd-ca184c65c38d" (UID: "db60029f-ec45-4c51-98cd-ca184c65c38d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.797171 4758 scope.go:117] "RemoveContainer" containerID="3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e" Dec 03 16:59:19 crc kubenswrapper[4758]: E1203 16:59:19.797753 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e\": container with ID starting with 3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e not found: ID does not exist" containerID="3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.797823 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e"} err="failed to get container status \"3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e\": rpc error: code = NotFound desc = could not find container \"3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e\": container with ID starting with 3db5ca4213650e175b873ecd05f0be0455e84b4125196346a6b5a4d1dd5e5c2e not found: ID does not exist" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.797866 4758 scope.go:117] "RemoveContainer" containerID="83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58" Dec 03 16:59:19 crc kubenswrapper[4758]: E1203 16:59:19.798257 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58\": container with ID starting with 83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58 not found: ID does not exist" containerID="83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.798296 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58"} err="failed to get container status \"83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58\": rpc error: code = NotFound desc = could not find container \"83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58\": container with ID starting with 83b397b6835c50bd6a82a04c0a6a4e6881f44d1c3bbff3c6ea3cc938f21d1e58 not found: ID does not exist" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.798327 4758 scope.go:117] "RemoveContainer" containerID="f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710" Dec 03 16:59:19 crc kubenswrapper[4758]: E1203 16:59:19.798548 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710\": container with ID starting with f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710 not found: ID does not exist" containerID="f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.798571 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710"} err="failed to get container status \"f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710\": rpc error: code = NotFound desc = could not find container \"f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710\": container with ID starting with f4bf25acedfa9a887584193d1744b884a1299ebb15fcb416ff821fe6c5940710 not found: ID does not exist" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.798585 4758 scope.go:117] "RemoveContainer" containerID="09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.819717 4758 scope.go:117] "RemoveContainer" containerID="b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.859543 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc530f3-6688-414d-9f8a-bb191ad5a264-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.859595 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db60029f-ec45-4c51-98cd-ca184c65c38d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.860957 4758 scope.go:117] "RemoveContainer" containerID="2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.886367 4758 scope.go:117] "RemoveContainer" containerID="09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268" Dec 03 16:59:19 crc kubenswrapper[4758]: E1203 16:59:19.887021 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268\": container with ID starting with 09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268 not found: ID does not exist" containerID="09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.887064 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268"} err="failed to get container status \"09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268\": rpc error: code = NotFound desc = could not find container \"09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268\": container with ID starting with 09aa8245c3b9175f537c29169f315aed2a26d1784d4b8f2e3b38fff906221268 not found: ID does not exist" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.887096 4758 scope.go:117] "RemoveContainer" containerID="b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8" Dec 03 16:59:19 crc kubenswrapper[4758]: E1203 16:59:19.887604 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8\": container with ID starting with b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8 not found: ID does not exist" containerID="b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.887625 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8"} err="failed to get container status \"b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8\": rpc error: code = NotFound desc = could not find container \"b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8\": container with ID starting with b2805847a9bcf403a70ec2c88978601f6433a5b08fe7f3e95674e108f0e58ce8 not found: ID does not exist" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.887639 4758 scope.go:117] "RemoveContainer" containerID="2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab" Dec 03 16:59:19 crc kubenswrapper[4758]: E1203 16:59:19.888240 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab\": container with ID starting with 2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab not found: ID does not exist" containerID="2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab" Dec 03 16:59:19 crc kubenswrapper[4758]: I1203 16:59:19.888262 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab"} err="failed to get container status \"2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab\": rpc error: code = NotFound desc = could not find container \"2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab\": container with ID starting with 2806d0d9949d499d072e6d6072dabb017d3962c78eec68256de5a62fd1a8faab not found: ID does not exist" Dec 03 16:59:20 crc kubenswrapper[4758]: I1203 16:59:20.040375 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jm64k"] Dec 03 16:59:20 crc kubenswrapper[4758]: I1203 16:59:20.046099 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jm64k"] Dec 03 16:59:20 crc kubenswrapper[4758]: I1203 16:59:20.058766 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l55d4"] Dec 03 16:59:20 crc kubenswrapper[4758]: I1203 16:59:20.062592 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l55d4"] Dec 03 16:59:21 crc kubenswrapper[4758]: I1203 16:59:21.123812 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" path="/var/lib/kubelet/pods/8dc530f3-6688-414d-9f8a-bb191ad5a264/volumes" Dec 03 16:59:21 crc kubenswrapper[4758]: I1203 16:59:21.124705 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" path="/var/lib/kubelet/pods/db60029f-ec45-4c51-98cd-ca184c65c38d/volumes" Dec 03 16:59:21 crc kubenswrapper[4758]: I1203 16:59:21.171219 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbcbx"] Dec 03 16:59:21 crc kubenswrapper[4758]: I1203 16:59:21.732768 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dbcbx" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="registry-server" containerID="cri-o://f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081" gracePeriod=2 Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.672816 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.741357 4758 generic.go:334] "Generic (PLEG): container finished" podID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerID="f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081" exitCode=0 Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.741404 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbcbx" event={"ID":"19fe802e-12eb-45e2-b918-ab7bb933dc04","Type":"ContainerDied","Data":"f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081"} Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.741432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbcbx" event={"ID":"19fe802e-12eb-45e2-b918-ab7bb933dc04","Type":"ContainerDied","Data":"77e04033a267a55c995daef9ca6439e34b203cda23f2632891c1b0f9d939c82e"} Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.741451 4758 scope.go:117] "RemoveContainer" containerID="f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.741545 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbcbx" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.764182 4758 scope.go:117] "RemoveContainer" containerID="9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.782751 4758 scope.go:117] "RemoveContainer" containerID="43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.796510 4758 scope.go:117] "RemoveContainer" containerID="f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081" Dec 03 16:59:22 crc kubenswrapper[4758]: E1203 16:59:22.796836 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081\": container with ID starting with f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081 not found: ID does not exist" containerID="f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.796864 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081"} err="failed to get container status \"f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081\": rpc error: code = NotFound desc = could not find container \"f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081\": container with ID starting with f7fae84569d56a5807b1cbaa7d587abeebdbcae405fa93958358b800042a3081 not found: ID does not exist" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.796886 4758 scope.go:117] "RemoveContainer" containerID="9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80" Dec 03 16:59:22 crc kubenswrapper[4758]: E1203 16:59:22.797154 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80\": container with ID starting with 9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80 not found: ID does not exist" containerID="9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.797178 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80"} err="failed to get container status \"9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80\": rpc error: code = NotFound desc = could not find container \"9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80\": container with ID starting with 9585f4c7459bcf0034a4c94586c09ad04b2868d3347215d128494bd99f427a80 not found: ID does not exist" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.797200 4758 scope.go:117] "RemoveContainer" containerID="43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a" Dec 03 16:59:22 crc kubenswrapper[4758]: E1203 16:59:22.797435 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a\": container with ID starting with 43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a not found: ID does not exist" containerID="43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.797466 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a"} err="failed to get container status \"43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a\": rpc error: code = NotFound desc = could not find container \"43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a\": container with ID starting with 43e4210ed054fe9ef1ac4a94f5a91e6c17013bf3d002f40eb44c56117b45d66a not found: ID does not exist" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.804776 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcsrz\" (UniqueName: \"kubernetes.io/projected/19fe802e-12eb-45e2-b918-ab7bb933dc04-kube-api-access-hcsrz\") pod \"19fe802e-12eb-45e2-b918-ab7bb933dc04\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.804825 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-utilities\") pod \"19fe802e-12eb-45e2-b918-ab7bb933dc04\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.804842 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-catalog-content\") pod \"19fe802e-12eb-45e2-b918-ab7bb933dc04\" (UID: \"19fe802e-12eb-45e2-b918-ab7bb933dc04\") " Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.807386 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-utilities" (OuterVolumeSpecName: "utilities") pod "19fe802e-12eb-45e2-b918-ab7bb933dc04" (UID: "19fe802e-12eb-45e2-b918-ab7bb933dc04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.810908 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19fe802e-12eb-45e2-b918-ab7bb933dc04-kube-api-access-hcsrz" (OuterVolumeSpecName: "kube-api-access-hcsrz") pod "19fe802e-12eb-45e2-b918-ab7bb933dc04" (UID: "19fe802e-12eb-45e2-b918-ab7bb933dc04"). InnerVolumeSpecName "kube-api-access-hcsrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.822358 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19fe802e-12eb-45e2-b918-ab7bb933dc04" (UID: "19fe802e-12eb-45e2-b918-ab7bb933dc04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.905275 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcsrz\" (UniqueName: \"kubernetes.io/projected/19fe802e-12eb-45e2-b918-ab7bb933dc04-kube-api-access-hcsrz\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.905308 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:22 crc kubenswrapper[4758]: I1203 16:59:22.905321 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19fe802e-12eb-45e2-b918-ab7bb933dc04-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.095707 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbcbx"] Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.101410 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbcbx"] Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.128036 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" path="/var/lib/kubelet/pods/19fe802e-12eb-45e2-b918-ab7bb933dc04/volumes" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.158990 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" podUID="71fd30bb-8da6-4b86-8dd6-56274220a60b" containerName="oauth-openshift" containerID="cri-o://358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510" gracePeriod=15 Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.640788 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.754794 4758 generic.go:334] "Generic (PLEG): container finished" podID="71fd30bb-8da6-4b86-8dd6-56274220a60b" containerID="358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510" exitCode=0 Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.754860 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" event={"ID":"71fd30bb-8da6-4b86-8dd6-56274220a60b","Type":"ContainerDied","Data":"358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510"} Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.754901 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" event={"ID":"71fd30bb-8da6-4b86-8dd6-56274220a60b","Type":"ContainerDied","Data":"6d479aa4ca446f444bd54948d86af20cab3ffb228cced346168c96d3796b0dbd"} Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.754936 4758 scope.go:117] "RemoveContainer" containerID="358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.754940 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nqrwv" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.797035 4758 scope.go:117] "RemoveContainer" containerID="358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.797860 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510\": container with ID starting with 358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510 not found: ID does not exist" containerID="358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.797928 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510"} err="failed to get container status \"358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510\": rpc error: code = NotFound desc = could not find container \"358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510\": container with ID starting with 358f44e242f9f7d9ec22e00ea7ee94759346fe39dab086f8deb058206ac67510 not found: ID does not exist" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821486 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-ocp-branding-template\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821537 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-router-certs\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821591 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-cliconfig\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821622 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-service-ca\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821654 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-policies\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821702 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-session\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821742 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-login\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821762 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-error\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821805 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-idp-0-file-data\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821826 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-serving-cert\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821848 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-trusted-ca-bundle\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821874 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-provider-selection\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821908 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b7qc\" (UniqueName: \"kubernetes.io/projected/71fd30bb-8da6-4b86-8dd6-56274220a60b-kube-api-access-9b7qc\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.821934 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-dir\") pod \"71fd30bb-8da6-4b86-8dd6-56274220a60b\" (UID: \"71fd30bb-8da6-4b86-8dd6-56274220a60b\") " Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.822671 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.823007 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.823747 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.823824 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.824334 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.829831 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.830582 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.831950 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71fd30bb-8da6-4b86-8dd6-56274220a60b-kube-api-access-9b7qc" (OuterVolumeSpecName: "kube-api-access-9b7qc") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "kube-api-access-9b7qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.832175 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.833162 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.833475 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.834043 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.834780 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.835851 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "71fd30bb-8da6-4b86-8dd6-56274220a60b" (UID: "71fd30bb-8da6-4b86-8dd6-56274220a60b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.903909 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg"] Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904254 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904282 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904317 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904335 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904358 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904375 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904393 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904408 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904430 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fd30bb-8da6-4b86-8dd6-56274220a60b" containerName="oauth-openshift" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904445 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fd30bb-8da6-4b86-8dd6-56274220a60b" containerName="oauth-openshift" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904501 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904517 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904537 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904551 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904572 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904586 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904601 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fdc199-9fb2-48bb-81a7-aab2454867eb" containerName="pruner" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904614 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fdc199-9fb2-48bb-81a7-aab2454867eb" containerName="pruner" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904634 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904647 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904669 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904723 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904746 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904759 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="extract-utilities" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904775 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904788 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="extract-content" Dec 03 16:59:23 crc kubenswrapper[4758]: E1203 16:59:23.904804 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904817 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.904992 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fdc199-9fb2-48bb-81a7-aab2454867eb" containerName="pruner" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.905018 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf1ea42-d625-4810-99f9-571f676ae801" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.905043 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="19fe802e-12eb-45e2-b918-ab7bb933dc04" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.905062 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="db60029f-ec45-4c51-98cd-ca184c65c38d" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.905082 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc530f3-6688-414d-9f8a-bb191ad5a264" containerName="registry-server" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.905095 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71fd30bb-8da6-4b86-8dd6-56274220a60b" containerName="oauth-openshift" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.905975 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.918161 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg"] Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.923908 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b7qc\" (UniqueName: \"kubernetes.io/projected/71fd30bb-8da6-4b86-8dd6-56274220a60b-kube-api-access-9b7qc\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.923957 4758 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.923981 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924003 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924024 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924042 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924060 4758 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924083 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924106 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924129 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924148 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924167 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924186 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:23 crc kubenswrapper[4758]: I1203 16:59:23.924212 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/71fd30bb-8da6-4b86-8dd6-56274220a60b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025330 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-audit-policies\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025396 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025421 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025466 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025493 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025529 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025665 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-login\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.025930 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e291041-164a-451d-82e7-d407c583fc39-audit-dir\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.026072 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-error\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.026138 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.026246 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l9zs\" (UniqueName: \"kubernetes.io/projected/7e291041-164a-451d-82e7-d407c583fc39-kube-api-access-4l9zs\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.026328 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-session\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.026398 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.026450 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.104874 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nqrwv"] Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.107814 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nqrwv"] Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-audit-policies\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128057 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128077 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128105 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128127 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128154 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128173 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-login\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128194 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e291041-164a-451d-82e7-d407c583fc39-audit-dir\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128214 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-error\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128230 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128247 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l9zs\" (UniqueName: \"kubernetes.io/projected/7e291041-164a-451d-82e7-d407c583fc39-kube-api-access-4l9zs\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128268 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-session\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128293 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.128309 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.129925 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e291041-164a-451d-82e7-d407c583fc39-audit-dir\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.130315 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.130894 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.131394 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.131604 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e291041-164a-451d-82e7-d407c583fc39-audit-policies\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.135457 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.136541 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-login\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.137729 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.142818 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-template-error\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.144494 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.145632 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.146015 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.149133 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7e291041-164a-451d-82e7-d407c583fc39-v4-0-config-system-session\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.153529 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l9zs\" (UniqueName: \"kubernetes.io/projected/7e291041-164a-451d-82e7-d407c583fc39-kube-api-access-4l9zs\") pod \"oauth-openshift-7c7b56dd96-s5nrg\" (UID: \"7e291041-164a-451d-82e7-d407c583fc39\") " pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.244467 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.507952 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg"] Dec 03 16:59:24 crc kubenswrapper[4758]: W1203 16:59:24.519412 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e291041_164a_451d_82e7_d407c583fc39.slice/crio-581b2e2afcba864d42c1040fdc5278e401ef3c951832ebd3ee50741b89c8c74e WatchSource:0}: Error finding container 581b2e2afcba864d42c1040fdc5278e401ef3c951832ebd3ee50741b89c8c74e: Status 404 returned error can't find the container with id 581b2e2afcba864d42c1040fdc5278e401ef3c951832ebd3ee50741b89c8c74e Dec 03 16:59:24 crc kubenswrapper[4758]: I1203 16:59:24.766665 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" event={"ID":"7e291041-164a-451d-82e7-d407c583fc39","Type":"ContainerStarted","Data":"581b2e2afcba864d42c1040fdc5278e401ef3c951832ebd3ee50741b89c8c74e"} Dec 03 16:59:25 crc kubenswrapper[4758]: I1203 16:59:25.131061 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71fd30bb-8da6-4b86-8dd6-56274220a60b" path="/var/lib/kubelet/pods/71fd30bb-8da6-4b86-8dd6-56274220a60b/volumes" Dec 03 16:59:25 crc kubenswrapper[4758]: I1203 16:59:25.776209 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" event={"ID":"7e291041-164a-451d-82e7-d407c583fc39","Type":"ContainerStarted","Data":"f6e6b5716fd2addb0377aa3961b962ae3eca5ab841fd8350cd09a16f18a34fd4"} Dec 03 16:59:25 crc kubenswrapper[4758]: I1203 16:59:25.777194 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:25 crc kubenswrapper[4758]: I1203 16:59:25.782965 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" Dec 03 16:59:25 crc kubenswrapper[4758]: I1203 16:59:25.815603 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7c7b56dd96-s5nrg" podStartSLOduration=27.815568309 podStartE2EDuration="27.815568309s" podCreationTimestamp="2025-12-03 16:58:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:59:25.811294013 +0000 UTC m=+221.012670934" watchObservedRunningTime="2025-12-03 16:59:25.815568309 +0000 UTC m=+221.016945210" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.435861 4758 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.437302 4758 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.437556 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425" gracePeriod=15 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.437723 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438055 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7" gracePeriod=15 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438113 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8" gracePeriod=15 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438208 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd" gracePeriod=15 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438288 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355" gracePeriod=15 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438556 4758 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438672 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438709 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438719 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438725 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438733 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438740 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438750 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438758 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438767 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438772 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438780 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438786 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.438798 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438806 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438897 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438908 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438917 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438928 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438938 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.438945 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.485557 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.595706 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.595866 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.595937 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.596070 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.596113 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.596134 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.596176 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.596207 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.697890 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.697951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.697971 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698001 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698030 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698049 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698067 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698095 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698170 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698200 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698214 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698236 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698267 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698201 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698298 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.698296 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.780393 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.805510 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.809014 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.811006 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd" exitCode=0 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.811064 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355" exitCode=0 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.811086 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7" exitCode=0 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.811109 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8" exitCode=2 Dec 03 16:59:28 crc kubenswrapper[4758]: I1203 16:59:28.811165 4758 scope.go:117] "RemoveContainer" containerID="4ac4aa01e3a69d0e50337dbf65039726ab853eb2b74e8e86641bb5f662b0b238" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.839769 4758 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc31f72d46d9d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:59:28.837029277 +0000 UTC m=+224.038406138,LastTimestamp:2025-12-03 16:59:28.837029277 +0000 UTC m=+224.038406138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.883838 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:28Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:28Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:28Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:28Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:27562b2c78cb890b90f8106f6f9225cc0da5a3232b1ac8a122bfc65917123a4e\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:62721bc9f3bcb671827417a74176173454e818cf8925f8bbe750618d2d94f8b6\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1609963837},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:446e5d504e70c7963ef7b0f090f3fcb19847ef48150299bf030847565d7a579b\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a01ee07f4838bab6cfa5a3d25d867557aa271725bfcd20a1e52d3cc63423c06b\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1204969293},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:31b0e25262b7daa1c7a43042f865ca936aa1a52776994642f88b9a12408d27da\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:ae694324b195581f542841a64634b63bae3d63332705b3a27320d18fde2aebe8\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201363276},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.884433 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.885582 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.886941 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.887815 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:28 crc kubenswrapper[4758]: E1203 16:59:28.887854 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.147041 4758 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.147720 4758 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.148137 4758 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.148600 4758 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.148828 4758 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:29 crc kubenswrapper[4758]: I1203 16:59:29.148856 4758 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.149116 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="200ms" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.350356 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="400ms" Dec 03 16:59:29 crc kubenswrapper[4758]: E1203 16:59:29.752465 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="800ms" Dec 03 16:59:29 crc kubenswrapper[4758]: I1203 16:59:29.821458 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01"} Dec 03 16:59:29 crc kubenswrapper[4758]: I1203 16:59:29.821532 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"41a2311e0c8bee7cf1a2a76109350fc03e8eadcec3d3bf88f45524106b885b90"} Dec 03 16:59:29 crc kubenswrapper[4758]: I1203 16:59:29.825021 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:59:29 crc kubenswrapper[4758]: I1203 16:59:29.826913 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:30 crc kubenswrapper[4758]: E1203 16:59:30.554151 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="1.6s" Dec 03 16:59:30 crc kubenswrapper[4758]: I1203 16:59:30.978809 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:59:30 crc kubenswrapper[4758]: I1203 16:59:30.980171 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:30 crc kubenswrapper[4758]: I1203 16:59:30.981478 4758 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:30 crc kubenswrapper[4758]: I1203 16:59:30.982312 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.130542 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.130707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.130781 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.130866 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.130928 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.131046 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.131393 4758 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.131438 4758 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.131457 4758 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:31 crc kubenswrapper[4758]: E1203 16:59:31.520728 4758 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc31f72d46d9d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:59:28.837029277 +0000 UTC m=+224.038406138,LastTimestamp:2025-12-03 16:59:28.837029277 +0000 UTC m=+224.038406138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.849348 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.851169 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425" exitCode=0 Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.851239 4758 scope.go:117] "RemoveContainer" containerID="4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.851270 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.853213 4758 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.853566 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.873848 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.874093 4758 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.884124 4758 scope.go:117] "RemoveContainer" containerID="fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.904529 4758 scope.go:117] "RemoveContainer" containerID="40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.920562 4758 scope.go:117] "RemoveContainer" containerID="f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.942576 4758 scope.go:117] "RemoveContainer" containerID="e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.967714 4758 scope.go:117] "RemoveContainer" containerID="db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.991473 4758 scope.go:117] "RemoveContainer" containerID="4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd" Dec 03 16:59:31 crc kubenswrapper[4758]: E1203 16:59:31.996532 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\": container with ID starting with 4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd not found: ID does not exist" containerID="4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.996581 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd"} err="failed to get container status \"4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\": rpc error: code = NotFound desc = could not find container \"4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd\": container with ID starting with 4b2aa98e221fa899516e54eb02254e3b50c06260c0fa7a3191c7828976b872dd not found: ID does not exist" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.996630 4758 scope.go:117] "RemoveContainer" containerID="fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355" Dec 03 16:59:31 crc kubenswrapper[4758]: E1203 16:59:31.997253 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\": container with ID starting with fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355 not found: ID does not exist" containerID="fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.997282 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355"} err="failed to get container status \"fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\": rpc error: code = NotFound desc = could not find container \"fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355\": container with ID starting with fec82d533b5062f7f297725ce03820205a476a74d166c7f34466ea75cdf55355 not found: ID does not exist" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.997301 4758 scope.go:117] "RemoveContainer" containerID="40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7" Dec 03 16:59:31 crc kubenswrapper[4758]: E1203 16:59:31.999096 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\": container with ID starting with 40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7 not found: ID does not exist" containerID="40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.999150 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7"} err="failed to get container status \"40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\": rpc error: code = NotFound desc = could not find container \"40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7\": container with ID starting with 40d576fdb098db9335300b2ade81ed075b30fb832e77d01693dff3a8ba4258d7 not found: ID does not exist" Dec 03 16:59:31 crc kubenswrapper[4758]: I1203 16:59:31.999189 4758 scope.go:117] "RemoveContainer" containerID="f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8" Dec 03 16:59:32 crc kubenswrapper[4758]: E1203 16:59:32.000292 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\": container with ID starting with f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8 not found: ID does not exist" containerID="f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8" Dec 03 16:59:32 crc kubenswrapper[4758]: I1203 16:59:32.000323 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8"} err="failed to get container status \"f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\": rpc error: code = NotFound desc = could not find container \"f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8\": container with ID starting with f4ee132e85c76a285030c9ac51dd88f2d726e11ea8aa43acc5daf6f6eb4f34d8 not found: ID does not exist" Dec 03 16:59:32 crc kubenswrapper[4758]: I1203 16:59:32.000338 4758 scope.go:117] "RemoveContainer" containerID="e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425" Dec 03 16:59:32 crc kubenswrapper[4758]: E1203 16:59:32.000889 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\": container with ID starting with e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425 not found: ID does not exist" containerID="e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425" Dec 03 16:59:32 crc kubenswrapper[4758]: I1203 16:59:32.000935 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425"} err="failed to get container status \"e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\": rpc error: code = NotFound desc = could not find container \"e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425\": container with ID starting with e8240ed1ff23a264f2eb6bba03a8c3f668147fab25edf26d41fb762ca7dfb425 not found: ID does not exist" Dec 03 16:59:32 crc kubenswrapper[4758]: I1203 16:59:32.000966 4758 scope.go:117] "RemoveContainer" containerID="db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c" Dec 03 16:59:32 crc kubenswrapper[4758]: E1203 16:59:32.001369 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\": container with ID starting with db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c not found: ID does not exist" containerID="db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c" Dec 03 16:59:32 crc kubenswrapper[4758]: I1203 16:59:32.001405 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c"} err="failed to get container status \"db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\": rpc error: code = NotFound desc = could not find container \"db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c\": container with ID starting with db09f643f58f18fa59664a6e14a2fd3760bad36e5ba9c08bafec3ca06bf9220c not found: ID does not exist" Dec 03 16:59:32 crc kubenswrapper[4758]: E1203 16:59:32.155614 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="3.2s" Dec 03 16:59:33 crc kubenswrapper[4758]: I1203 16:59:33.121656 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 16:59:33 crc kubenswrapper[4758]: I1203 16:59:33.873440 4758 generic.go:334] "Generic (PLEG): container finished" podID="0e7787c1-2384-45d9-a620-096dd4ed60e0" containerID="dc23826b9bf8bc185f2a8a0189b5c11746f7f048e0d4ed78eba5925e711c954a" exitCode=0 Dec 03 16:59:33 crc kubenswrapper[4758]: I1203 16:59:33.873500 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0e7787c1-2384-45d9-a620-096dd4ed60e0","Type":"ContainerDied","Data":"dc23826b9bf8bc185f2a8a0189b5c11746f7f048e0d4ed78eba5925e711c954a"} Dec 03 16:59:33 crc kubenswrapper[4758]: I1203 16:59:33.874289 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:33 crc kubenswrapper[4758]: I1203 16:59:33.874590 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.117157 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.117804 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.147457 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.148176 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.148729 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.284776 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e7787c1-2384-45d9-a620-096dd4ed60e0-kube-api-access\") pod \"0e7787c1-2384-45d9-a620-096dd4ed60e0\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.284855 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-kubelet-dir\") pod \"0e7787c1-2384-45d9-a620-096dd4ed60e0\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.285009 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0e7787c1-2384-45d9-a620-096dd4ed60e0" (UID: "0e7787c1-2384-45d9-a620-096dd4ed60e0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.284989 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-var-lock\") pod \"0e7787c1-2384-45d9-a620-096dd4ed60e0\" (UID: \"0e7787c1-2384-45d9-a620-096dd4ed60e0\") " Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.285088 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-var-lock" (OuterVolumeSpecName: "var-lock") pod "0e7787c1-2384-45d9-a620-096dd4ed60e0" (UID: "0e7787c1-2384-45d9-a620-096dd4ed60e0"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.285761 4758 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.285783 4758 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0e7787c1-2384-45d9-a620-096dd4ed60e0-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.291846 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e7787c1-2384-45d9-a620-096dd4ed60e0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0e7787c1-2384-45d9-a620-096dd4ed60e0" (UID: "0e7787c1-2384-45d9-a620-096dd4ed60e0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:59:35 crc kubenswrapper[4758]: E1203 16:59:35.357178 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="6.4s" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.386570 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e7787c1-2384-45d9-a620-096dd4ed60e0-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.887461 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0e7787c1-2384-45d9-a620-096dd4ed60e0","Type":"ContainerDied","Data":"d69e27ca93ec2a6c1020af8c2a29d4821eff3f82719baf5aaa0092a669034157"} Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.887857 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d69e27ca93ec2a6c1020af8c2a29d4821eff3f82719baf5aaa0092a669034157" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.887721 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.901862 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:35 crc kubenswrapper[4758]: I1203 16:59:35.902374 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:39 crc kubenswrapper[4758]: E1203 16:59:39.139927 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:39Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:39Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:39Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:59:39Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:27562b2c78cb890b90f8106f6f9225cc0da5a3232b1ac8a122bfc65917123a4e\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:62721bc9f3bcb671827417a74176173454e818cf8925f8bbe750618d2d94f8b6\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1609963837},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:446e5d504e70c7963ef7b0f090f3fcb19847ef48150299bf030847565d7a579b\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a01ee07f4838bab6cfa5a3d25d867557aa271725bfcd20a1e52d3cc63423c06b\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1204969293},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:31b0e25262b7daa1c7a43042f865ca936aa1a52776994642f88b9a12408d27da\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:ae694324b195581f542841a64634b63bae3d63332705b3a27320d18fde2aebe8\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201363276},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:39 crc kubenswrapper[4758]: E1203 16:59:39.140790 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:39 crc kubenswrapper[4758]: E1203 16:59:39.141111 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:39 crc kubenswrapper[4758]: E1203 16:59:39.141352 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:39 crc kubenswrapper[4758]: E1203 16:59:39.141580 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:39 crc kubenswrapper[4758]: E1203 16:59:39.141598 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:59:41 crc kubenswrapper[4758]: E1203 16:59:41.522975 4758 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc31f72d46d9d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:59:28.837029277 +0000 UTC m=+224.038406138,LastTimestamp:2025-12-03 16:59:28.837029277 +0000 UTC m=+224.038406138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:59:41 crc kubenswrapper[4758]: E1203 16:59:41.758887 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="7s" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.114289 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.115493 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.117563 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.157235 4758 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.157281 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:42 crc kubenswrapper[4758]: E1203 16:59:42.157850 4758 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.158527 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:42 crc kubenswrapper[4758]: W1203 16:59:42.192346 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-533afbd0c602ea0eaf31527205cfdc0dfc0efacf64f3cf65e6df394ad635cadb WatchSource:0}: Error finding container 533afbd0c602ea0eaf31527205cfdc0dfc0efacf64f3cf65e6df394ad635cadb: Status 404 returned error can't find the container with id 533afbd0c602ea0eaf31527205cfdc0dfc0efacf64f3cf65e6df394ad635cadb Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.937619 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.938212 4758 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644" exitCode=1 Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.938289 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644"} Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.939134 4758 scope.go:117] "RemoveContainer" containerID="761adc7ab12b5f76b2914e815912ed92e4735346dcfe28b67a064d40fa96b644" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.939397 4758 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.939597 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.940015 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.943637 4758 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="a1433866796ac34b32e30cb98da0e7239df91cb2c649f3794ac2f22bf6706ce8" exitCode=0 Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.943722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"a1433866796ac34b32e30cb98da0e7239df91cb2c649f3794ac2f22bf6706ce8"} Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.943780 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"533afbd0c602ea0eaf31527205cfdc0dfc0efacf64f3cf65e6df394ad635cadb"} Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.944147 4758 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.944173 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:42 crc kubenswrapper[4758]: E1203 16:59:42.944694 4758 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.944953 4758 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.945259 4758 status_manager.go:851] "Failed to get status for pod" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:42 crc kubenswrapper[4758]: I1203 16:59:42.945770 4758 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.720978 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.954219 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b5eb1f6905d1ce5ee896ae6fdc2b0106ba374c643845b183a88fdfead6df8c1a"} Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.954267 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"48fe4b81b186213142c1af46935aafb2856bc6b433a2e9fc997db32a2482ce21"} Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.954280 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c14bbc357bdc84f22059fe6c114d704fe02e5373e6f4438ac68506fcd2e44e17"} Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.954290 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e801b11f7a563b520133167025e0e2250e03322540604d4400cf2189450055ad"} Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.957773 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 16:59:43 crc kubenswrapper[4758]: I1203 16:59:43.957855 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a788bde7832209c071d00c091f7c8eb016cf330c27463f06c736bdc044631f5a"} Dec 03 16:59:44 crc kubenswrapper[4758]: I1203 16:59:44.966323 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"900d1a4ed99332b793b4ed7db36789cac6239644e7ebdb83ee68e255c24e3d01"} Dec 03 16:59:44 crc kubenswrapper[4758]: I1203 16:59:44.966705 4758 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:44 crc kubenswrapper[4758]: I1203 16:59:44.966732 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:46 crc kubenswrapper[4758]: I1203 16:59:46.124919 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:59:46 crc kubenswrapper[4758]: I1203 16:59:46.129504 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:59:46 crc kubenswrapper[4758]: I1203 16:59:46.980775 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:59:47 crc kubenswrapper[4758]: I1203 16:59:47.159468 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:47 crc kubenswrapper[4758]: I1203 16:59:47.159896 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:47 crc kubenswrapper[4758]: I1203 16:59:47.165540 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:49 crc kubenswrapper[4758]: I1203 16:59:49.978470 4758 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:50 crc kubenswrapper[4758]: I1203 16:59:50.004710 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:50 crc kubenswrapper[4758]: I1203 16:59:50.004724 4758 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:50 crc kubenswrapper[4758]: I1203 16:59:50.004748 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:50 crc kubenswrapper[4758]: I1203 16:59:50.008856 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:59:50 crc kubenswrapper[4758]: I1203 16:59:50.011175 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="13ed5890-2211-46c8-b68a-67aaa092dc17" Dec 03 16:59:51 crc kubenswrapper[4758]: I1203 16:59:51.014979 4758 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:51 crc kubenswrapper[4758]: I1203 16:59:51.015035 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:51 crc kubenswrapper[4758]: I1203 16:59:51.018600 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="13ed5890-2211-46c8-b68a-67aaa092dc17" Dec 03 16:59:52 crc kubenswrapper[4758]: I1203 16:59:52.019534 4758 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:52 crc kubenswrapper[4758]: I1203 16:59:52.019588 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="69900819-66e0-4891-bd32-cd06492eea18" Dec 03 16:59:52 crc kubenswrapper[4758]: I1203 16:59:52.022767 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="13ed5890-2211-46c8-b68a-67aaa092dc17" Dec 03 16:59:53 crc kubenswrapper[4758]: I1203 16:59:53.734482 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:59:59 crc kubenswrapper[4758]: I1203 16:59:59.738168 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.273998 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.374816 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.426060 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.498164 4758 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.589091 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.622355 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.660097 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 17:00:00 crc kubenswrapper[4758]: I1203 17:00:00.815218 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:00:01 crc kubenswrapper[4758]: I1203 17:00:01.046506 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 17:00:01 crc kubenswrapper[4758]: I1203 17:00:01.249469 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 17:00:01 crc kubenswrapper[4758]: I1203 17:00:01.384116 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 17:00:01 crc kubenswrapper[4758]: I1203 17:00:01.797787 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 17:00:01 crc kubenswrapper[4758]: I1203 17:00:01.946376 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.167884 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.569852 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.662734 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.719216 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.757776 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.796272 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 17:00:02 crc kubenswrapper[4758]: I1203 17:00:02.972337 4758 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.055467 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.261484 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.262025 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.267553 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.332826 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.374816 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.388714 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.514129 4758 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.617605 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.644365 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.702623 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.729976 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.746335 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.795745 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.879730 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.969830 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 17:00:03 crc kubenswrapper[4758]: I1203 17:00:03.996748 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.009341 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.066443 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.067090 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.074164 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.159267 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.253960 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.258846 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.269871 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.294743 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.315880 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.332552 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.369844 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.529571 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.660055 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.669028 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.789609 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.807715 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.872506 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.889700 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.984067 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 17:00:04 crc kubenswrapper[4758]: I1203 17:00:04.992769 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.025325 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.096093 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.203790 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.238741 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.265534 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.277264 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.296054 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.396714 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.467245 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.471158 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.588999 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.688789 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.728428 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.731662 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.742445 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.824373 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.834110 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.868201 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.895957 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.907568 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.936530 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 17:00:05 crc kubenswrapper[4758]: I1203 17:00:05.955991 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.011719 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.012907 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.038982 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.082358 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.109136 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.157584 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.260761 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.328328 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.435065 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.456395 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.663670 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.700798 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.811919 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.862404 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.885733 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.926558 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 17:00:06 crc kubenswrapper[4758]: I1203 17:00:06.995101 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.074822 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.103984 4758 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.112549 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.306325 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.360797 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.381767 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.403336 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.414515 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.451723 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.526842 4758 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.528497 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=39.528464056 podStartE2EDuration="39.528464056s" podCreationTimestamp="2025-12-03 16:59:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:59:49.905973789 +0000 UTC m=+245.107350670" watchObservedRunningTime="2025-12-03 17:00:07.528464056 +0000 UTC m=+262.729840957" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.535732 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.535811 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.541549 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.558025 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.557997046 podStartE2EDuration="18.557997046s" podCreationTimestamp="2025-12-03 16:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:00:07.555785587 +0000 UTC m=+262.757162438" watchObservedRunningTime="2025-12-03 17:00:07.557997046 +0000 UTC m=+262.759373947" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.593608 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.605912 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.691932 4758 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.786655 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.793759 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.856801 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.900622 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.925982 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 17:00:07 crc kubenswrapper[4758]: I1203 17:00:07.993146 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.052963 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.060998 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.077156 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.188232 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.235197 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.236962 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.269720 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.328107 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.454171 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.462136 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.486007 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.496381 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.518313 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.602004 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.612585 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.860152 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.884574 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 17:00:08 crc kubenswrapper[4758]: I1203 17:00:08.886748 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.028603 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.052288 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.172871 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.262904 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.336758 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.440085 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.446312 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.464039 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.465471 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.465839 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.468924 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.539537 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.551277 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.617401 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.617401 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.680000 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.798379 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 17:00:09 crc kubenswrapper[4758]: I1203 17:00:09.868393 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.016610 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.038076 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.275923 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.289345 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.322441 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.331124 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.541032 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.585155 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.589572 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.598129 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.607991 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.623067 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.634502 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.788028 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.821820 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.860284 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.874655 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.901878 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.924254 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.945123 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 17:00:10 crc kubenswrapper[4758]: I1203 17:00:10.983767 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.053010 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.065103 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.114948 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.119655 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.147865 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.201459 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.203605 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.203926 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.206460 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.244579 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.254721 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.281062 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.281173 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.309999 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.333735 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.334501 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.355160 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.386479 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.510575 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.511347 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.516963 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.538453 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.558147 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.599123 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.656594 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.777464 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.877798 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.961804 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.963499 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:00:11 crc kubenswrapper[4758]: I1203 17:00:11.977734 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.050242 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.097475 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.320342 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.325778 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.352886 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.364649 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.409102 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf"] Dec 03 17:00:12 crc kubenswrapper[4758]: E1203 17:00:12.409613 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" containerName="installer" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.409702 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" containerName="installer" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.409875 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e7787c1-2384-45d9-a620-096dd4ed60e0" containerName="installer" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.410329 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.416661 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.417766 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.423641 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf"] Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.524169 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a25891-06ed-4b49-ab23-6551b6903199-secret-volume\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.524228 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a25891-06ed-4b49-ab23-6551b6903199-config-volume\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.524353 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvxl\" (UniqueName: \"kubernetes.io/projected/58a25891-06ed-4b49-ab23-6551b6903199-kube-api-access-wsvxl\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.560364 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.625468 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a25891-06ed-4b49-ab23-6551b6903199-secret-volume\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.625548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a25891-06ed-4b49-ab23-6551b6903199-config-volume\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.625570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvxl\" (UniqueName: \"kubernetes.io/projected/58a25891-06ed-4b49-ab23-6551b6903199-kube-api-access-wsvxl\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.626823 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a25891-06ed-4b49-ab23-6551b6903199-config-volume\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.642810 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a25891-06ed-4b49-ab23-6551b6903199-secret-volume\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.644712 4758 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.645029 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01" gracePeriod=5 Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.657245 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvxl\" (UniqueName: \"kubernetes.io/projected/58a25891-06ed-4b49-ab23-6551b6903199-kube-api-access-wsvxl\") pod \"collect-profiles-29413020-ck5lf\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.722107 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.730586 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.769439 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.860303 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.881052 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.902393 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 17:00:12 crc kubenswrapper[4758]: I1203 17:00:12.986931 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.043200 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.059496 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf"] Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.152419 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" event={"ID":"58a25891-06ed-4b49-ab23-6551b6903199","Type":"ContainerStarted","Data":"e3eb3eaa0455840632f8da509cea0b4c470cfb84d9f08eede7140ac26e1f3550"} Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.221863 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.272956 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.336882 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.398950 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.520960 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.676458 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.677076 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.690923 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.709032 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.766643 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 17:00:13 crc kubenswrapper[4758]: I1203 17:00:13.861570 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.110134 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.129359 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.159985 4758 generic.go:334] "Generic (PLEG): container finished" podID="58a25891-06ed-4b49-ab23-6551b6903199" containerID="53a2bbd9d7f93ea90625056446dfd2ea459065e55c56c4f570c4d2ce7cf5ee74" exitCode=0 Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.160033 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" event={"ID":"58a25891-06ed-4b49-ab23-6551b6903199","Type":"ContainerDied","Data":"53a2bbd9d7f93ea90625056446dfd2ea459065e55c56c4f570c4d2ce7cf5ee74"} Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.194331 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.221988 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.289965 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4b9m"] Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.290260 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d4b9m" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="registry-server" containerID="cri-o://72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8" gracePeriod=30 Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.295780 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7kgk7"] Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.296115 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7kgk7" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="registry-server" containerID="cri-o://cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518" gracePeriod=30 Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.313499 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.315403 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8sqq"] Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.315778 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" containerID="cri-o://531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd" gracePeriod=30 Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.324349 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5wnr"] Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.324822 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r5wnr" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="registry-server" containerID="cri-o://7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d" gracePeriod=30 Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.342186 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8qlms"] Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.342477 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8qlms" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="registry-server" containerID="cri-o://fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7" gracePeriod=30 Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.357715 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fgbb5"] Dec 03 17:00:14 crc kubenswrapper[4758]: E1203 17:00:14.358035 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.358056 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.358189 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.358849 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.359112 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.376379 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fgbb5"] Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.456872 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.552795 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmk8j\" (UniqueName: \"kubernetes.io/projected/6a4c5544-6114-40e4-8d89-f1a60bc1708a-kube-api-access-qmk8j\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.552880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6a4c5544-6114-40e4-8d89-f1a60bc1708a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.552960 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a4c5544-6114-40e4-8d89-f1a60bc1708a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.594838 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.595293 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.608218 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.655039 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmk8j\" (UniqueName: \"kubernetes.io/projected/6a4c5544-6114-40e4-8d89-f1a60bc1708a-kube-api-access-qmk8j\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.655115 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6a4c5544-6114-40e4-8d89-f1a60bc1708a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.655217 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a4c5544-6114-40e4-8d89-f1a60bc1708a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.657786 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a4c5544-6114-40e4-8d89-f1a60bc1708a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.665971 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6a4c5544-6114-40e4-8d89-f1a60bc1708a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.673362 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmk8j\" (UniqueName: \"kubernetes.io/projected/6a4c5544-6114-40e4-8d89-f1a60bc1708a-kube-api-access-qmk8j\") pod \"marketplace-operator-79b997595-fgbb5\" (UID: \"6a4c5544-6114-40e4-8d89-f1a60bc1708a\") " pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.771494 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.777745 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863056 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863507 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-utilities\") pod \"b43b0669-6da9-438d-bc36-90456cc2611c\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-catalog-content\") pod \"f673f32d-f81f-4fe9-ab86-801fe50a46da\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863615 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-utilities\") pod \"f673f32d-f81f-4fe9-ab86-801fe50a46da\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863631 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-catalog-content\") pod \"b43b0669-6da9-438d-bc36-90456cc2611c\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863718 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms74c\" (UniqueName: \"kubernetes.io/projected/b43b0669-6da9-438d-bc36-90456cc2611c-kube-api-access-ms74c\") pod \"b43b0669-6da9-438d-bc36-90456cc2611c\" (UID: \"b43b0669-6da9-438d-bc36-90456cc2611c\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.863744 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnbjk\" (UniqueName: \"kubernetes.io/projected/f673f32d-f81f-4fe9-ab86-801fe50a46da-kube-api-access-dnbjk\") pod \"f673f32d-f81f-4fe9-ab86-801fe50a46da\" (UID: \"f673f32d-f81f-4fe9-ab86-801fe50a46da\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.864837 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-utilities" (OuterVolumeSpecName: "utilities") pod "b43b0669-6da9-438d-bc36-90456cc2611c" (UID: "b43b0669-6da9-438d-bc36-90456cc2611c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.870940 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f673f32d-f81f-4fe9-ab86-801fe50a46da-kube-api-access-dnbjk" (OuterVolumeSpecName: "kube-api-access-dnbjk") pod "f673f32d-f81f-4fe9-ab86-801fe50a46da" (UID: "f673f32d-f81f-4fe9-ab86-801fe50a46da"). InnerVolumeSpecName "kube-api-access-dnbjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.871387 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-utilities" (OuterVolumeSpecName: "utilities") pod "f673f32d-f81f-4fe9-ab86-801fe50a46da" (UID: "f673f32d-f81f-4fe9-ab86-801fe50a46da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.871783 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43b0669-6da9-438d-bc36-90456cc2611c-kube-api-access-ms74c" (OuterVolumeSpecName: "kube-api-access-ms74c") pod "b43b0669-6da9-438d-bc36-90456cc2611c" (UID: "b43b0669-6da9-438d-bc36-90456cc2611c"). InnerVolumeSpecName "kube-api-access-ms74c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.896291 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.931553 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964298 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-catalog-content\") pod \"d1afe03b-b381-4fc5-973a-f8178760890e\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-utilities\") pod \"d1afe03b-b381-4fc5-973a-f8178760890e\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964376 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv2xt\" (UniqueName: \"kubernetes.io/projected/d1afe03b-b381-4fc5-973a-f8178760890e-kube-api-access-dv2xt\") pod \"d1afe03b-b381-4fc5-973a-f8178760890e\" (UID: \"d1afe03b-b381-4fc5-973a-f8178760890e\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964396 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv9cj\" (UniqueName: \"kubernetes.io/projected/83593597-7147-4b61-be7c-e05b165bdebe-kube-api-access-zv9cj\") pod \"83593597-7147-4b61-be7c-e05b165bdebe\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964419 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-trusted-ca\") pod \"83593597-7147-4b61-be7c-e05b165bdebe\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964447 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-operator-metrics\") pod \"83593597-7147-4b61-be7c-e05b165bdebe\" (UID: \"83593597-7147-4b61-be7c-e05b165bdebe\") " Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964626 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964640 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms74c\" (UniqueName: \"kubernetes.io/projected/b43b0669-6da9-438d-bc36-90456cc2611c-kube-api-access-ms74c\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964650 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnbjk\" (UniqueName: \"kubernetes.io/projected/f673f32d-f81f-4fe9-ab86-801fe50a46da-kube-api-access-dnbjk\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.964659 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.968493 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "83593597-7147-4b61-be7c-e05b165bdebe" (UID: "83593597-7147-4b61-be7c-e05b165bdebe"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.968934 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1afe03b-b381-4fc5-973a-f8178760890e-kube-api-access-dv2xt" (OuterVolumeSpecName: "kube-api-access-dv2xt") pod "d1afe03b-b381-4fc5-973a-f8178760890e" (UID: "d1afe03b-b381-4fc5-973a-f8178760890e"). InnerVolumeSpecName "kube-api-access-dv2xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.968964 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "83593597-7147-4b61-be7c-e05b165bdebe" (UID: "83593597-7147-4b61-be7c-e05b165bdebe"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.969394 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-utilities" (OuterVolumeSpecName: "utilities") pod "d1afe03b-b381-4fc5-973a-f8178760890e" (UID: "d1afe03b-b381-4fc5-973a-f8178760890e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.975005 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83593597-7147-4b61-be7c-e05b165bdebe-kube-api-access-zv9cj" (OuterVolumeSpecName: "kube-api-access-zv9cj") pod "83593597-7147-4b61-be7c-e05b165bdebe" (UID: "83593597-7147-4b61-be7c-e05b165bdebe"). InnerVolumeSpecName "kube-api-access-zv9cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.980964 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.982088 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f673f32d-f81f-4fe9-ab86-801fe50a46da" (UID: "f673f32d-f81f-4fe9-ab86-801fe50a46da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.985719 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b43b0669-6da9-438d-bc36-90456cc2611c" (UID: "b43b0669-6da9-438d-bc36-90456cc2611c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:14 crc kubenswrapper[4758]: I1203 17:00:14.995778 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1afe03b-b381-4fc5-973a-f8178760890e" (UID: "d1afe03b-b381-4fc5-973a-f8178760890e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.024794 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.065853 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066130 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f673f32d-f81f-4fe9-ab86-801fe50a46da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066226 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b43b0669-6da9-438d-bc36-90456cc2611c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066303 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066377 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1afe03b-b381-4fc5-973a-f8178760890e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066454 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv2xt\" (UniqueName: \"kubernetes.io/projected/d1afe03b-b381-4fc5-973a-f8178760890e-kube-api-access-dv2xt\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066527 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv9cj\" (UniqueName: \"kubernetes.io/projected/83593597-7147-4b61-be7c-e05b165bdebe-kube-api-access-zv9cj\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.066616 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83593597-7147-4b61-be7c-e05b165bdebe-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.077846 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fgbb5"] Dec 03 17:00:15 crc kubenswrapper[4758]: W1203 17:00:15.091427 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a4c5544_6114_40e4_8d89_f1a60bc1708a.slice/crio-00c320912e14b3fba570c2accc623430fbb03d53e6cb588b6fbe77984709c402 WatchSource:0}: Error finding container 00c320912e14b3fba570c2accc623430fbb03d53e6cb588b6fbe77984709c402: Status 404 returned error can't find the container with id 00c320912e14b3fba570c2accc623430fbb03d53e6cb588b6fbe77984709c402 Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.121862 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.167491 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-catalog-content\") pod \"8c7f12da-f734-4617-8977-3004953eb1aa\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.167845 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-utilities\") pod \"8c7f12da-f734-4617-8977-3004953eb1aa\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.167931 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5pkc\" (UniqueName: \"kubernetes.io/projected/8c7f12da-f734-4617-8977-3004953eb1aa-kube-api-access-j5pkc\") pod \"8c7f12da-f734-4617-8977-3004953eb1aa\" (UID: \"8c7f12da-f734-4617-8977-3004953eb1aa\") " Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.171252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-utilities" (OuterVolumeSpecName: "utilities") pod "8c7f12da-f734-4617-8977-3004953eb1aa" (UID: "8c7f12da-f734-4617-8977-3004953eb1aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.179881 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7f12da-f734-4617-8977-3004953eb1aa-kube-api-access-j5pkc" (OuterVolumeSpecName: "kube-api-access-j5pkc") pod "8c7f12da-f734-4617-8977-3004953eb1aa" (UID: "8c7f12da-f734-4617-8977-3004953eb1aa"). InnerVolumeSpecName "kube-api-access-j5pkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.186422 4758 generic.go:334] "Generic (PLEG): container finished" podID="83593597-7147-4b61-be7c-e05b165bdebe" containerID="531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd" exitCode=0 Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.187028 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.187224 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" event={"ID":"83593597-7147-4b61-be7c-e05b165bdebe","Type":"ContainerDied","Data":"531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.187322 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r8sqq" event={"ID":"83593597-7147-4b61-be7c-e05b165bdebe","Type":"ContainerDied","Data":"128fc95a96bc235f23f52262de8749879b7e5255b470a20ef23180daf4dcfdd8"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.187345 4758 scope.go:117] "RemoveContainer" containerID="531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.190297 4758 generic.go:334] "Generic (PLEG): container finished" podID="b43b0669-6da9-438d-bc36-90456cc2611c" containerID="72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8" exitCode=0 Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.190410 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4b9m" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.190497 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerDied","Data":"72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.190573 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4b9m" event={"ID":"b43b0669-6da9-438d-bc36-90456cc2611c","Type":"ContainerDied","Data":"a87273ce688f26abfd57001490d4479b2df9df8458882a8455945015b385b72a"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.192337 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" event={"ID":"6a4c5544-6114-40e4-8d89-f1a60bc1708a","Type":"ContainerStarted","Data":"00c320912e14b3fba570c2accc623430fbb03d53e6cb588b6fbe77984709c402"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.196153 4758 generic.go:334] "Generic (PLEG): container finished" podID="8c7f12da-f734-4617-8977-3004953eb1aa" containerID="fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7" exitCode=0 Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.196256 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qlms" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.196273 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerDied","Data":"fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.196315 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qlms" event={"ID":"8c7f12da-f734-4617-8977-3004953eb1aa","Type":"ContainerDied","Data":"02b8265c5330f87dbc7e7c5a7e2796160342abba85eeda497088873adbe064a6"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.200113 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1afe03b-b381-4fc5-973a-f8178760890e" containerID="7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d" exitCode=0 Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.200157 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerDied","Data":"7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.200229 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5wnr" event={"ID":"d1afe03b-b381-4fc5-973a-f8178760890e","Type":"ContainerDied","Data":"3fc1d4112b366056dc67b50e3efde4ffe77b0f6163d05a5e3946affc76a369f9"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.201171 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r5wnr" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.213781 4758 scope.go:117] "RemoveContainer" containerID="531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.214890 4758 generic.go:334] "Generic (PLEG): container finished" podID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerID="cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518" exitCode=0 Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.215199 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerDied","Data":"cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.215259 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7kgk7" event={"ID":"f673f32d-f81f-4fe9-ab86-801fe50a46da","Type":"ContainerDied","Data":"78da1b48c47d2d2c2619e246a1e57b2e7fcd0edd4f28ac353159315b3b3781a0"} Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.215285 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7kgk7" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.215594 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd\": container with ID starting with 531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd not found: ID does not exist" containerID="531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.215643 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd"} err="failed to get container status \"531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd\": rpc error: code = NotFound desc = could not find container \"531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd\": container with ID starting with 531d1649ac16165db4b5dda3c75f70eb4ad2d25d272e70c3988ee02a016bc9bd not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.216383 4758 scope.go:117] "RemoveContainer" containerID="72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.229140 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8sqq"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.239005 4758 scope.go:117] "RemoveContainer" containerID="2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.239485 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r8sqq"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.245049 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4b9m"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.253325 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d4b9m"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.262284 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7kgk7"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.266223 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7kgk7"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.270924 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.270979 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5pkc\" (UniqueName: \"kubernetes.io/projected/8c7f12da-f734-4617-8977-3004953eb1aa-kube-api-access-j5pkc\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.273778 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5wnr"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.275797 4758 scope.go:117] "RemoveContainer" containerID="49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.278634 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5wnr"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.300784 4758 scope.go:117] "RemoveContainer" containerID="72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.301716 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8\": container with ID starting with 72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8 not found: ID does not exist" containerID="72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.301753 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8"} err="failed to get container status \"72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8\": rpc error: code = NotFound desc = could not find container \"72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8\": container with ID starting with 72473fde68a52ac439374ba6e986db77a4f9594341c58a791d43c716215da3c8 not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.301781 4758 scope.go:117] "RemoveContainer" containerID="2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.302057 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf\": container with ID starting with 2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf not found: ID does not exist" containerID="2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.302080 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf"} err="failed to get container status \"2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf\": rpc error: code = NotFound desc = could not find container \"2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf\": container with ID starting with 2f17871d3c8c58a7223544476c31ea68fd3edaf2d9d7cd1036a0ab8fccaa5ddf not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.302099 4758 scope.go:117] "RemoveContainer" containerID="49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.302482 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf\": container with ID starting with 49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf not found: ID does not exist" containerID="49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.302516 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf"} err="failed to get container status \"49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf\": rpc error: code = NotFound desc = could not find container \"49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf\": container with ID starting with 49f55089d07a1f11e060d03f8957b7e685a7cf65fbc316f4a449562a7fb2feaf not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.302534 4758 scope.go:117] "RemoveContainer" containerID="fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.328309 4758 scope.go:117] "RemoveContainer" containerID="6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.360072 4758 scope.go:117] "RemoveContainer" containerID="2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.361550 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c7f12da-f734-4617-8977-3004953eb1aa" (UID: "8c7f12da-f734-4617-8977-3004953eb1aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.372354 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7f12da-f734-4617-8977-3004953eb1aa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.396665 4758 scope.go:117] "RemoveContainer" containerID="fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.399243 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7\": container with ID starting with fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7 not found: ID does not exist" containerID="fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.399279 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7"} err="failed to get container status \"fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7\": rpc error: code = NotFound desc = could not find container \"fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7\": container with ID starting with fefe46aad88b56989c8e8c96be3fcb3d9294ce6f3b318e1e6f6e30a1fb0057d7 not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.399307 4758 scope.go:117] "RemoveContainer" containerID="6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.399599 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e\": container with ID starting with 6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e not found: ID does not exist" containerID="6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.399628 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e"} err="failed to get container status \"6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e\": rpc error: code = NotFound desc = could not find container \"6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e\": container with ID starting with 6ac1108d02354a969b9d3b10d79c821875dea1e8969599bb5675f999571b8e8e not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.399649 4758 scope.go:117] "RemoveContainer" containerID="2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.399887 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d\": container with ID starting with 2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d not found: ID does not exist" containerID="2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.399915 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d"} err="failed to get container status \"2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d\": rpc error: code = NotFound desc = could not find container \"2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d\": container with ID starting with 2a02c38e274794c7cc2e8f969b085be8e99692fda8b944dc081f38355ef96e9d not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.399932 4758 scope.go:117] "RemoveContainer" containerID="7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.420976 4758 scope.go:117] "RemoveContainer" containerID="fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.436996 4758 scope.go:117] "RemoveContainer" containerID="5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.437303 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.452129 4758 scope.go:117] "RemoveContainer" containerID="7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.453078 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d\": container with ID starting with 7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d not found: ID does not exist" containerID="7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.453120 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d"} err="failed to get container status \"7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d\": rpc error: code = NotFound desc = could not find container \"7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d\": container with ID starting with 7b13d37bd98e3e2fa492ac4ca7e9db49532ed04d51f6b44d2a581a36aae1e67d not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.453147 4758 scope.go:117] "RemoveContainer" containerID="fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.454357 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43\": container with ID starting with fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43 not found: ID does not exist" containerID="fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.454385 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43"} err="failed to get container status \"fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43\": rpc error: code = NotFound desc = could not find container \"fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43\": container with ID starting with fb66e42617bde0007987570b32f5f015c6577245a47b8ffcd5d5c549cc3fac43 not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.454412 4758 scope.go:117] "RemoveContainer" containerID="5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.454642 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa\": container with ID starting with 5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa not found: ID does not exist" containerID="5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.454662 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa"} err="failed to get container status \"5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa\": rpc error: code = NotFound desc = could not find container \"5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa\": container with ID starting with 5189eb53322e160218ea3d3a578e0470d136671cb8abde890806d20290431eaa not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.454722 4758 scope.go:117] "RemoveContainer" containerID="cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.471546 4758 scope.go:117] "RemoveContainer" containerID="18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.472650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsvxl\" (UniqueName: \"kubernetes.io/projected/58a25891-06ed-4b49-ab23-6551b6903199-kube-api-access-wsvxl\") pod \"58a25891-06ed-4b49-ab23-6551b6903199\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.472739 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a25891-06ed-4b49-ab23-6551b6903199-config-volume\") pod \"58a25891-06ed-4b49-ab23-6551b6903199\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.472814 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a25891-06ed-4b49-ab23-6551b6903199-secret-volume\") pod \"58a25891-06ed-4b49-ab23-6551b6903199\" (UID: \"58a25891-06ed-4b49-ab23-6551b6903199\") " Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.473712 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58a25891-06ed-4b49-ab23-6551b6903199-config-volume" (OuterVolumeSpecName: "config-volume") pod "58a25891-06ed-4b49-ab23-6551b6903199" (UID: "58a25891-06ed-4b49-ab23-6551b6903199"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.478466 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a25891-06ed-4b49-ab23-6551b6903199-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "58a25891-06ed-4b49-ab23-6551b6903199" (UID: "58a25891-06ed-4b49-ab23-6551b6903199"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.478495 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a25891-06ed-4b49-ab23-6551b6903199-kube-api-access-wsvxl" (OuterVolumeSpecName: "kube-api-access-wsvxl") pod "58a25891-06ed-4b49-ab23-6551b6903199" (UID: "58a25891-06ed-4b49-ab23-6551b6903199"). InnerVolumeSpecName "kube-api-access-wsvxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.493869 4758 scope.go:117] "RemoveContainer" containerID="382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.517803 4758 scope.go:117] "RemoveContainer" containerID="cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.518198 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518\": container with ID starting with cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518 not found: ID does not exist" containerID="cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.518240 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518"} err="failed to get container status \"cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518\": rpc error: code = NotFound desc = could not find container \"cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518\": container with ID starting with cae34f0285366a0287f322a908c32873d4e8a8fb54608b4738105ce0b5443518 not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.518276 4758 scope.go:117] "RemoveContainer" containerID="18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.519190 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b\": container with ID starting with 18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b not found: ID does not exist" containerID="18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.519235 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b"} err="failed to get container status \"18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b\": rpc error: code = NotFound desc = could not find container \"18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b\": container with ID starting with 18b5b45d1ee94bf4bd34132ad13d8e6f3a6a50281370dd7824ebe454ff1b564b not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.519262 4758 scope.go:117] "RemoveContainer" containerID="382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86" Dec 03 17:00:15 crc kubenswrapper[4758]: E1203 17:00:15.520434 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86\": container with ID starting with 382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86 not found: ID does not exist" containerID="382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.520458 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86"} err="failed to get container status \"382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86\": rpc error: code = NotFound desc = could not find container \"382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86\": container with ID starting with 382939b3f8db30e40b7f99defe40ef390c1398d24c3dc56d9392c8e59d643a86 not found: ID does not exist" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.534001 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8qlms"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.539368 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8qlms"] Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.574129 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a25891-06ed-4b49-ab23-6551b6903199-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.574169 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a25891-06ed-4b49-ab23-6551b6903199-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.574185 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsvxl\" (UniqueName: \"kubernetes.io/projected/58a25891-06ed-4b49-ab23-6551b6903199-kube-api-access-wsvxl\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.644737 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.883075 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.915219 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.949251 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 17:00:15 crc kubenswrapper[4758]: I1203 17:00:15.988599 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.052536 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.071468 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.234725 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" event={"ID":"58a25891-06ed-4b49-ab23-6551b6903199","Type":"ContainerDied","Data":"e3eb3eaa0455840632f8da509cea0b4c470cfb84d9f08eede7140ac26e1f3550"} Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.234812 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3eb3eaa0455840632f8da509cea0b4c470cfb84d9f08eede7140ac26e1f3550" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.234828 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.237375 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" event={"ID":"6a4c5544-6114-40e4-8d89-f1a60bc1708a","Type":"ContainerStarted","Data":"5cf0b332d4be0fc5d35df5d86e82a993af7eafe1e687f33bccf51e77ccf78db0"} Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.238129 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.248131 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.271478 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fgbb5" podStartSLOduration=2.271447553 podStartE2EDuration="2.271447553s" podCreationTimestamp="2025-12-03 17:00:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:00:16.268233586 +0000 UTC m=+271.469610487" watchObservedRunningTime="2025-12-03 17:00:16.271447553 +0000 UTC m=+271.472824484" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.349965 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.456639 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:00:16 crc kubenswrapper[4758]: I1203 17:00:16.548015 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 17:00:17 crc kubenswrapper[4758]: I1203 17:00:17.133997 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83593597-7147-4b61-be7c-e05b165bdebe" path="/var/lib/kubelet/pods/83593597-7147-4b61-be7c-e05b165bdebe/volumes" Dec 03 17:00:17 crc kubenswrapper[4758]: I1203 17:00:17.134849 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" path="/var/lib/kubelet/pods/8c7f12da-f734-4617-8977-3004953eb1aa/volumes" Dec 03 17:00:17 crc kubenswrapper[4758]: I1203 17:00:17.136558 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" path="/var/lib/kubelet/pods/b43b0669-6da9-438d-bc36-90456cc2611c/volumes" Dec 03 17:00:17 crc kubenswrapper[4758]: I1203 17:00:17.138565 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" path="/var/lib/kubelet/pods/d1afe03b-b381-4fc5-973a-f8178760890e/volumes" Dec 03 17:00:17 crc kubenswrapper[4758]: I1203 17:00:17.139813 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" path="/var/lib/kubelet/pods/f673f32d-f81f-4fe9-ab86-801fe50a46da/volumes" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.237056 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.237601 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.250927 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.251040 4758 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01" exitCode=137 Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.251119 4758 scope.go:117] "RemoveContainer" containerID="e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.251168 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.275058 4758 scope.go:117] "RemoveContainer" containerID="e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01" Dec 03 17:00:18 crc kubenswrapper[4758]: E1203 17:00:18.275878 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01\": container with ID starting with e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01 not found: ID does not exist" containerID="e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.275921 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01"} err="failed to get container status \"e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01\": rpc error: code = NotFound desc = could not find container \"e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01\": container with ID starting with e456afc804d1e14bfca59df7ef0e8888dbe051db69f36fd085b1d75825aa2b01 not found: ID does not exist" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418480 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418713 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418743 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418774 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418806 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418827 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418848 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.419336 4758 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.419376 4758 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.419394 4758 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.418861 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.436866 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.520226 4758 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.520297 4758 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:18 crc kubenswrapper[4758]: I1203 17:00:18.657242 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 17:00:19 crc kubenswrapper[4758]: I1203 17:00:19.121969 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 17:00:19 crc kubenswrapper[4758]: I1203 17:00:19.122780 4758 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 03 17:00:19 crc kubenswrapper[4758]: I1203 17:00:19.133251 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 17:00:19 crc kubenswrapper[4758]: I1203 17:00:19.133320 4758 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="441e837f-744e-4aee-9c43-cae2ad7226a9" Dec 03 17:00:19 crc kubenswrapper[4758]: I1203 17:00:19.136217 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 17:00:19 crc kubenswrapper[4758]: I1203 17:00:19.136264 4758 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="441e837f-744e-4aee-9c43-cae2ad7226a9" Dec 03 17:00:47 crc kubenswrapper[4758]: I1203 17:00:47.542885 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4f6xq"] Dec 03 17:00:47 crc kubenswrapper[4758]: I1203 17:00:47.543804 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" podUID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" containerName="controller-manager" containerID="cri-o://533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f" gracePeriod=30 Dec 03 17:00:47 crc kubenswrapper[4758]: I1203 17:00:47.597314 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7"] Dec 03 17:00:47 crc kubenswrapper[4758]: I1203 17:00:47.597525 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" podUID="c30fe471-6fb2-467a-b63e-2fc8f915c948" containerName="route-controller-manager" containerID="cri-o://97c83131a914a6b8fcba7cf698bed9360d69706ddbf60f03419e618bc55482f2" gracePeriod=30 Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.464013 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.470237 4758 generic.go:334] "Generic (PLEG): container finished" podID="c30fe471-6fb2-467a-b63e-2fc8f915c948" containerID="97c83131a914a6b8fcba7cf698bed9360d69706ddbf60f03419e618bc55482f2" exitCode=0 Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.470292 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" event={"ID":"c30fe471-6fb2-467a-b63e-2fc8f915c948","Type":"ContainerDied","Data":"97c83131a914a6b8fcba7cf698bed9360d69706ddbf60f03419e618bc55482f2"} Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.472114 4758 generic.go:334] "Generic (PLEG): container finished" podID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" containerID="533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f" exitCode=0 Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.472138 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" event={"ID":"71308db0-b934-4c92-b90c-aaf1c6b1b41c","Type":"ContainerDied","Data":"533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f"} Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.472152 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" event={"ID":"71308db0-b934-4c92-b90c-aaf1c6b1b41c","Type":"ContainerDied","Data":"cc50c38f4bc54d617e83b3d0384acdfc21f0d4f922adc1a7eb60e6ed7c3e279e"} Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.472164 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4f6xq" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.472186 4758 scope.go:117] "RemoveContainer" containerID="533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.496638 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71308db0-b934-4c92-b90c-aaf1c6b1b41c-serving-cert\") pod \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.496736 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-config\") pod \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.496777 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-proxy-ca-bundles\") pod \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.496801 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-client-ca\") pod \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.496820 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7psj4\" (UniqueName: \"kubernetes.io/projected/71308db0-b934-4c92-b90c-aaf1c6b1b41c-kube-api-access-7psj4\") pod \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\" (UID: \"71308db0-b934-4c92-b90c-aaf1c6b1b41c\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.497938 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "71308db0-b934-4c92-b90c-aaf1c6b1b41c" (UID: "71308db0-b934-4c92-b90c-aaf1c6b1b41c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.497948 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-client-ca" (OuterVolumeSpecName: "client-ca") pod "71308db0-b934-4c92-b90c-aaf1c6b1b41c" (UID: "71308db0-b934-4c92-b90c-aaf1c6b1b41c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.498021 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-config" (OuterVolumeSpecName: "config") pod "71308db0-b934-4c92-b90c-aaf1c6b1b41c" (UID: "71308db0-b934-4c92-b90c-aaf1c6b1b41c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.498625 4758 scope.go:117] "RemoveContainer" containerID="533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f" Dec 03 17:00:48 crc kubenswrapper[4758]: E1203 17:00:48.499710 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f\": container with ID starting with 533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f not found: ID does not exist" containerID="533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.499762 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f"} err="failed to get container status \"533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f\": rpc error: code = NotFound desc = could not find container \"533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f\": container with ID starting with 533a41c8c74eed421f6ef50a68e5295646678d595fe9003baf977fcf7d28f94f not found: ID does not exist" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.504320 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71308db0-b934-4c92-b90c-aaf1c6b1b41c-kube-api-access-7psj4" (OuterVolumeSpecName: "kube-api-access-7psj4") pod "71308db0-b934-4c92-b90c-aaf1c6b1b41c" (UID: "71308db0-b934-4c92-b90c-aaf1c6b1b41c"). InnerVolumeSpecName "kube-api-access-7psj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.504920 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71308db0-b934-4c92-b90c-aaf1c6b1b41c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "71308db0-b934-4c92-b90c-aaf1c6b1b41c" (UID: "71308db0-b934-4c92-b90c-aaf1c6b1b41c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.514151 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.597834 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctqp9\" (UniqueName: \"kubernetes.io/projected/c30fe471-6fb2-467a-b63e-2fc8f915c948-kube-api-access-ctqp9\") pod \"c30fe471-6fb2-467a-b63e-2fc8f915c948\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.597896 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-client-ca\") pod \"c30fe471-6fb2-467a-b63e-2fc8f915c948\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.597983 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c30fe471-6fb2-467a-b63e-2fc8f915c948-serving-cert\") pod \"c30fe471-6fb2-467a-b63e-2fc8f915c948\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.598027 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-config\") pod \"c30fe471-6fb2-467a-b63e-2fc8f915c948\" (UID: \"c30fe471-6fb2-467a-b63e-2fc8f915c948\") " Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.598214 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71308db0-b934-4c92-b90c-aaf1c6b1b41c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.598252 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.598266 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.598284 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7psj4\" (UniqueName: \"kubernetes.io/projected/71308db0-b934-4c92-b90c-aaf1c6b1b41c-kube-api-access-7psj4\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.598297 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71308db0-b934-4c92-b90c-aaf1c6b1b41c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.599317 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-config" (OuterVolumeSpecName: "config") pod "c30fe471-6fb2-467a-b63e-2fc8f915c948" (UID: "c30fe471-6fb2-467a-b63e-2fc8f915c948"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.599383 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-client-ca" (OuterVolumeSpecName: "client-ca") pod "c30fe471-6fb2-467a-b63e-2fc8f915c948" (UID: "c30fe471-6fb2-467a-b63e-2fc8f915c948"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.602142 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c30fe471-6fb2-467a-b63e-2fc8f915c948-kube-api-access-ctqp9" (OuterVolumeSpecName: "kube-api-access-ctqp9") pod "c30fe471-6fb2-467a-b63e-2fc8f915c948" (UID: "c30fe471-6fb2-467a-b63e-2fc8f915c948"). InnerVolumeSpecName "kube-api-access-ctqp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.602549 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c30fe471-6fb2-467a-b63e-2fc8f915c948-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c30fe471-6fb2-467a-b63e-2fc8f915c948" (UID: "c30fe471-6fb2-467a-b63e-2fc8f915c948"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.699332 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c30fe471-6fb2-467a-b63e-2fc8f915c948-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.699377 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.699391 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctqp9\" (UniqueName: \"kubernetes.io/projected/c30fe471-6fb2-467a-b63e-2fc8f915c948-kube-api-access-ctqp9\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.699407 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c30fe471-6fb2-467a-b63e-2fc8f915c948-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.796706 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4f6xq"] Dec 03 17:00:48 crc kubenswrapper[4758]: I1203 17:00:48.803706 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4f6xq"] Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.126321 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" path="/var/lib/kubelet/pods/71308db0-b934-4c92-b90c-aaf1c6b1b41c/volumes" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.484407 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" event={"ID":"c30fe471-6fb2-467a-b63e-2fc8f915c948","Type":"ContainerDied","Data":"668a66256a32352e1ce33cafacd211bd718d0255aea36e696887b4b2f9c33cdb"} Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.484847 4758 scope.go:117] "RemoveContainer" containerID="97c83131a914a6b8fcba7cf698bed9360d69706ddbf60f03419e618bc55482f2" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.484511 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.510041 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7"] Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.516035 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm7"] Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.963440 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r"] Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.963805 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.963827 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.963847 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.963860 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.963885 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.963899 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.963915 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.963980 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964000 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" containerName="controller-manager" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964014 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" containerName="controller-manager" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964033 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964048 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964068 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964081 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964100 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a25891-06ed-4b49-ab23-6551b6903199" containerName="collect-profiles" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964112 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a25891-06ed-4b49-ab23-6551b6903199" containerName="collect-profiles" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964132 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964145 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="extract-content" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964161 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964173 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964191 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964204 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964227 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30fe471-6fb2-467a-b63e-2fc8f915c948" containerName="route-controller-manager" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964239 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30fe471-6fb2-467a-b63e-2fc8f915c948" containerName="route-controller-manager" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964256 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964269 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964289 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964302 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964319 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964331 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" Dec 03 17:00:49 crc kubenswrapper[4758]: E1203 17:00:49.964350 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964364 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="extract-utilities" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964520 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f673f32d-f81f-4fe9-ab86-801fe50a46da" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964543 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afe03b-b381-4fc5-973a-f8178760890e" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964566 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43b0669-6da9-438d-bc36-90456cc2611c" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964583 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c30fe471-6fb2-467a-b63e-2fc8f915c948" containerName="route-controller-manager" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964602 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="83593597-7147-4b61-be7c-e05b165bdebe" containerName="marketplace-operator" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964617 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a25891-06ed-4b49-ab23-6551b6903199" containerName="collect-profiles" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964633 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71308db0-b934-4c92-b90c-aaf1c6b1b41c" containerName="controller-manager" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.964653 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7f12da-f734-4617-8977-3004953eb1aa" containerName="registry-server" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.965233 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.968677 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.968980 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.970075 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.971173 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.971277 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.971480 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.976024 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-8nrzz"] Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.977831 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.982994 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.983458 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.983885 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.984026 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.984106 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.983877 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:00:49 crc kubenswrapper[4758]: I1203 17:00:49.988849 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-8nrzz"] Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.000200 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r"] Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.000361 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.114827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-config\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.114971 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr854\" (UniqueName: \"kubernetes.io/projected/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-kube-api-access-wr854\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.115914 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c7f556-8630-4320-a7fc-9d7a16f06bb8-serving-cert\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.116088 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-serving-cert\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.116131 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-client-ca\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.116235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8969\" (UniqueName: \"kubernetes.io/projected/13c7f556-8630-4320-a7fc-9d7a16f06bb8-kube-api-access-k8969\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.116323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-proxy-ca-bundles\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.116466 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-config\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.116506 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-client-ca\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.217914 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr854\" (UniqueName: \"kubernetes.io/projected/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-kube-api-access-wr854\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.217981 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c7f556-8630-4320-a7fc-9d7a16f06bb8-serving-cert\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218015 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-serving-cert\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-client-ca\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218057 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8969\" (UniqueName: \"kubernetes.io/projected/13c7f556-8630-4320-a7fc-9d7a16f06bb8-kube-api-access-k8969\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218081 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-proxy-ca-bundles\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218107 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-config\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218120 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-client-ca\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.218144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-config\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.219836 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-config\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.220043 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-client-ca\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.220263 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-client-ca\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.220373 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-proxy-ca-bundles\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.220752 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-config\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.224179 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c7f556-8630-4320-a7fc-9d7a16f06bb8-serving-cert\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.225534 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-serving-cert\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.248130 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr854\" (UniqueName: \"kubernetes.io/projected/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-kube-api-access-wr854\") pod \"controller-manager-5747cbd54d-8nrzz\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.251050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8969\" (UniqueName: \"kubernetes.io/projected/13c7f556-8630-4320-a7fc-9d7a16f06bb8-kube-api-access-k8969\") pod \"route-controller-manager-5f7f4579db-pp55r\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.292646 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.310219 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.513313 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-8nrzz"] Dec 03 17:00:50 crc kubenswrapper[4758]: I1203 17:00:50.557462 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r"] Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.132845 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c30fe471-6fb2-467a-b63e-2fc8f915c948" path="/var/lib/kubelet/pods/c30fe471-6fb2-467a-b63e-2fc8f915c948/volumes" Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.504818 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" event={"ID":"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca","Type":"ContainerStarted","Data":"9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e"} Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.505726 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" event={"ID":"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca","Type":"ContainerStarted","Data":"346444ce47f1e5c14eda5969581e6400a8baba9c1f8eeef2234044814407925a"} Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.506201 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.507833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" event={"ID":"13c7f556-8630-4320-a7fc-9d7a16f06bb8","Type":"ContainerStarted","Data":"5839ac89eab3962554481ddafdfa4ea8510f1d7d7328afa186422a3c63c97883"} Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.508609 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.508738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" event={"ID":"13c7f556-8630-4320-a7fc-9d7a16f06bb8","Type":"ContainerStarted","Data":"29223a90819738d8c7335c2c5d2088b3e9e9fcfe90d8aaadb3e08685ffadbaa4"} Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.514812 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.536506 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" podStartSLOduration=4.536472849 podStartE2EDuration="4.536472849s" podCreationTimestamp="2025-12-03 17:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:00:51.526357435 +0000 UTC m=+306.727734316" watchObservedRunningTime="2025-12-03 17:00:51.536472849 +0000 UTC m=+306.737849750" Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.538621 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:00:51 crc kubenswrapper[4758]: I1203 17:00:51.551735 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" podStartSLOduration=3.551702992 podStartE2EDuration="3.551702992s" podCreationTimestamp="2025-12-03 17:00:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:00:51.550828529 +0000 UTC m=+306.752205400" watchObservedRunningTime="2025-12-03 17:00:51.551702992 +0000 UTC m=+306.753079853" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.572784 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t52js"] Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.575156 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.581598 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.583159 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t52js"] Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.732750 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-catalog-content\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.732814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg5zx\" (UniqueName: \"kubernetes.io/projected/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-kube-api-access-rg5zx\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.732846 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-utilities\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.770007 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m8ltr"] Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.771961 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.774518 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.799462 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m8ltr"] Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.834499 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-catalog-content\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.835182 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-catalog-content\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.835447 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg5zx\" (UniqueName: \"kubernetes.io/projected/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-kube-api-access-rg5zx\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.835798 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-utilities\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.836163 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-utilities\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.860875 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bdk6p"] Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.862036 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.866104 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg5zx\" (UniqueName: \"kubernetes.io/projected/6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6-kube-api-access-rg5zx\") pod \"redhat-marketplace-t52js\" (UID: \"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6\") " pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.879253 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bdk6p"] Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.914113 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.937247 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d77c9\" (UniqueName: \"kubernetes.io/projected/45bb5bc6-b9e2-403f-9811-8223b2b660cb-kube-api-access-d77c9\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.937312 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-catalog-content\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:05 crc kubenswrapper[4758]: I1203 17:01:05.937473 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-utilities\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039350 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/47e81ea5-4b47-404a-8639-aea46b41c78b-registry-certificates\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039450 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-catalog-content\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039497 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47e81ea5-4b47-404a-8639-aea46b41c78b-trusted-ca\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039543 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/47e81ea5-4b47-404a-8639-aea46b41c78b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/47e81ea5-4b47-404a-8639-aea46b41c78b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039612 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-utilities\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039701 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-bound-sa-token\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039758 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039818 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwf74\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-kube-api-access-mwf74\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039865 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-registry-tls\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.039916 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d77c9\" (UniqueName: \"kubernetes.io/projected/45bb5bc6-b9e2-403f-9811-8223b2b660cb-kube-api-access-d77c9\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.040024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-catalog-content\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.040322 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-utilities\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.061909 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d77c9\" (UniqueName: \"kubernetes.io/projected/45bb5bc6-b9e2-403f-9811-8223b2b660cb-kube-api-access-d77c9\") pod \"redhat-operators-m8ltr\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.067901 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.095180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142694 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/47e81ea5-4b47-404a-8639-aea46b41c78b-registry-certificates\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142762 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47e81ea5-4b47-404a-8639-aea46b41c78b-trusted-ca\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142822 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/47e81ea5-4b47-404a-8639-aea46b41c78b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142841 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/47e81ea5-4b47-404a-8639-aea46b41c78b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142916 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-bound-sa-token\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142958 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwf74\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-kube-api-access-mwf74\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.142983 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-registry-tls\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.144010 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/47e81ea5-4b47-404a-8639-aea46b41c78b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.144417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/47e81ea5-4b47-404a-8639-aea46b41c78b-registry-certificates\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.145057 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47e81ea5-4b47-404a-8639-aea46b41c78b-trusted-ca\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.147556 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/47e81ea5-4b47-404a-8639-aea46b41c78b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.147624 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-registry-tls\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.165333 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-bound-sa-token\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.174132 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwf74\" (UniqueName: \"kubernetes.io/projected/47e81ea5-4b47-404a-8639-aea46b41c78b-kube-api-access-mwf74\") pod \"image-registry-66df7c8f76-bdk6p\" (UID: \"47e81ea5-4b47-404a-8639-aea46b41c78b\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.205758 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.312391 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t52js"] Dec 03 17:01:06 crc kubenswrapper[4758]: W1203 17:01:06.322899 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d9c8c5c_e835_4f95_8c0d_7e6c7cd1f1a6.slice/crio-7084125749ce65623f828f405d25a8dd0e19e982040e352ada99ab3922a28b2d WatchSource:0}: Error finding container 7084125749ce65623f828f405d25a8dd0e19e982040e352ada99ab3922a28b2d: Status 404 returned error can't find the container with id 7084125749ce65623f828f405d25a8dd0e19e982040e352ada99ab3922a28b2d Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.524296 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m8ltr"] Dec 03 17:01:06 crc kubenswrapper[4758]: W1203 17:01:06.528062 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45bb5bc6_b9e2_403f_9811_8223b2b660cb.slice/crio-46e73544c10d2de2db2df61b8f6efd3fe6688b06843de1e3349a2bce467aab23 WatchSource:0}: Error finding container 46e73544c10d2de2db2df61b8f6efd3fe6688b06843de1e3349a2bce467aab23: Status 404 returned error can't find the container with id 46e73544c10d2de2db2df61b8f6efd3fe6688b06843de1e3349a2bce467aab23 Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.609520 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8ltr" event={"ID":"45bb5bc6-b9e2-403f-9811-8223b2b660cb","Type":"ContainerStarted","Data":"46e73544c10d2de2db2df61b8f6efd3fe6688b06843de1e3349a2bce467aab23"} Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.614980 4758 generic.go:334] "Generic (PLEG): container finished" podID="6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6" containerID="b6401b2ac159a4d83a82a57b641463e69cea591f743b86c7fbe19c6116dbea2c" exitCode=0 Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.615072 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t52js" event={"ID":"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6","Type":"ContainerDied","Data":"b6401b2ac159a4d83a82a57b641463e69cea591f743b86c7fbe19c6116dbea2c"} Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.615137 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t52js" event={"ID":"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6","Type":"ContainerStarted","Data":"7084125749ce65623f828f405d25a8dd0e19e982040e352ada99ab3922a28b2d"} Dec 03 17:01:06 crc kubenswrapper[4758]: I1203 17:01:06.655386 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bdk6p"] Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.580028 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-8nrzz"] Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.580371 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" podUID="a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" containerName="controller-manager" containerID="cri-o://9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e" gracePeriod=30 Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.623753 4758 generic.go:334] "Generic (PLEG): container finished" podID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerID="8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3" exitCode=0 Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.623858 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8ltr" event={"ID":"45bb5bc6-b9e2-403f-9811-8223b2b660cb","Type":"ContainerDied","Data":"8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3"} Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.625480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" event={"ID":"47e81ea5-4b47-404a-8639-aea46b41c78b","Type":"ContainerStarted","Data":"43433b196dc8ac83dfd4a1429202c896d5a9b04c5aa489ea23d525883905c50b"} Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.625988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" event={"ID":"47e81ea5-4b47-404a-8639-aea46b41c78b","Type":"ContainerStarted","Data":"200f43c92186f98160e64e409886f4b55982ae4669f9f9396e2c0cf1501f59f1"} Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.626019 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.629594 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t52js" event={"ID":"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6","Type":"ContainerStarted","Data":"2b16aff2e06d5d360d61cc405a8eed69bd562f6d58bf50cf7aa2f2c890615f41"} Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.684740 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" podStartSLOduration=2.684716131 podStartE2EDuration="2.684716131s" podCreationTimestamp="2025-12-03 17:01:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:01:07.683160799 +0000 UTC m=+322.884537670" watchObservedRunningTime="2025-12-03 17:01:07.684716131 +0000 UTC m=+322.886092982" Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.976188 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-spff6"] Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.977531 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.981356 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 17:01:07 crc kubenswrapper[4758]: I1203 17:01:07.990852 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-spff6"] Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.159887 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.165323 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-26n6q"] Dec 03 17:01:08 crc kubenswrapper[4758]: E1203 17:01:08.165596 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" containerName="controller-manager" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.165616 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" containerName="controller-manager" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.165745 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" containerName="controller-manager" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.166516 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.169046 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.172042 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-utilities\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.172092 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-kube-api-access-l26gl\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.172138 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-catalog-content\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.177736 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-26n6q"] Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.273645 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-config\") pod \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.273741 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr854\" (UniqueName: \"kubernetes.io/projected/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-kube-api-access-wr854\") pod \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.273806 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-client-ca\") pod \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.273844 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-proxy-ca-bundles\") pod \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.273871 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-serving-cert\") pod \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\" (UID: \"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca\") " Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274200 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-kube-api-access-l26gl\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274254 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a247927-7fef-4538-b8bb-d3b2d146e318-utilities\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274287 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-catalog-content\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8pmf\" (UniqueName: \"kubernetes.io/projected/3a247927-7fef-4538-b8bb-d3b2d146e318-kube-api-access-q8pmf\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274360 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a247927-7fef-4538-b8bb-d3b2d146e318-catalog-content\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274392 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-utilities\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.274941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-utilities\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.275643 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-client-ca" (OuterVolumeSpecName: "client-ca") pod "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" (UID: "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.275634 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" (UID: "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.275934 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-catalog-content\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.276501 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-config" (OuterVolumeSpecName: "config") pod "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" (UID: "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.284211 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" (UID: "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.284328 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-kube-api-access-wr854" (OuterVolumeSpecName: "kube-api-access-wr854") pod "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" (UID: "a2bebe1e-fa8c-4a55-a89b-5d4d662672ca"). InnerVolumeSpecName "kube-api-access-wr854". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.300623 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f-kube-api-access-l26gl\") pod \"community-operators-spff6\" (UID: \"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f\") " pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.375873 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a247927-7fef-4538-b8bb-d3b2d146e318-utilities\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.375968 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8pmf\" (UniqueName: \"kubernetes.io/projected/3a247927-7fef-4538-b8bb-d3b2d146e318-kube-api-access-q8pmf\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376011 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a247927-7fef-4538-b8bb-d3b2d146e318-catalog-content\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376102 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376118 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376130 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376146 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376155 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr854\" (UniqueName: \"kubernetes.io/projected/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca-kube-api-access-wr854\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.376665 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a247927-7fef-4538-b8bb-d3b2d146e318-catalog-content\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.377187 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a247927-7fef-4538-b8bb-d3b2d146e318-utilities\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.393888 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8pmf\" (UniqueName: \"kubernetes.io/projected/3a247927-7fef-4538-b8bb-d3b2d146e318-kube-api-access-q8pmf\") pod \"certified-operators-26n6q\" (UID: \"3a247927-7fef-4538-b8bb-d3b2d146e318\") " pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.484440 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.612622 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.641308 4758 generic.go:334] "Generic (PLEG): container finished" podID="a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" containerID="9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e" exitCode=0 Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.641396 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" event={"ID":"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca","Type":"ContainerDied","Data":"9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e"} Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.641432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" event={"ID":"a2bebe1e-fa8c-4a55-a89b-5d4d662672ca","Type":"ContainerDied","Data":"346444ce47f1e5c14eda5969581e6400a8baba9c1f8eeef2234044814407925a"} Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.641456 4758 scope.go:117] "RemoveContainer" containerID="9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.641609 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-8nrzz" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.646364 4758 generic.go:334] "Generic (PLEG): container finished" podID="6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6" containerID="2b16aff2e06d5d360d61cc405a8eed69bd562f6d58bf50cf7aa2f2c890615f41" exitCode=0 Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.646416 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t52js" event={"ID":"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6","Type":"ContainerDied","Data":"2b16aff2e06d5d360d61cc405a8eed69bd562f6d58bf50cf7aa2f2c890615f41"} Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.666309 4758 scope.go:117] "RemoveContainer" containerID="9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e" Dec 03 17:01:08 crc kubenswrapper[4758]: E1203 17:01:08.667146 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e\": container with ID starting with 9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e not found: ID does not exist" containerID="9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.667196 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e"} err="failed to get container status \"9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e\": rpc error: code = NotFound desc = could not find container \"9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e\": container with ID starting with 9d03900455b39ff2d2109f30aaa06a54cd41ece47a6d09f98dc1f686d3ead26e not found: ID does not exist" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.705345 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-8nrzz"] Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.739324 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-8nrzz"] Dec 03 17:01:08 crc kubenswrapper[4758]: E1203 17:01:08.788834 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2bebe1e_fa8c_4a55_a89b_5d4d662672ca.slice/crio-346444ce47f1e5c14eda5969581e6400a8baba9c1f8eeef2234044814407925a\": RecentStats: unable to find data in memory cache]" Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.880337 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-spff6"] Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.940106 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-26n6q"] Dec 03 17:01:08 crc kubenswrapper[4758]: I1203 17:01:08.984072 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d76f496d8-xcjb6"] Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:08.998103 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.007604 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.007855 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.008009 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.008183 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.008353 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.020804 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.021313 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.024624 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d76f496d8-xcjb6"] Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.090836 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5tk4\" (UniqueName: \"kubernetes.io/projected/31568899-a9c0-438f-b273-c54f9ab55ffb-kube-api-access-f5tk4\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.090902 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-client-ca\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.090932 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-config\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.090964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31568899-a9c0-438f-b273-c54f9ab55ffb-serving-cert\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.091000 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-proxy-ca-bundles\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.122226 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2bebe1e-fa8c-4a55-a89b-5d4d662672ca" path="/var/lib/kubelet/pods/a2bebe1e-fa8c-4a55-a89b-5d4d662672ca/volumes" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.192640 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-config\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.192749 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31568899-a9c0-438f-b273-c54f9ab55ffb-serving-cert\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.192810 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-proxy-ca-bundles\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.192862 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5tk4\" (UniqueName: \"kubernetes.io/projected/31568899-a9c0-438f-b273-c54f9ab55ffb-kube-api-access-f5tk4\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.192893 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-client-ca\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.194610 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-config\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.194835 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-client-ca\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.196149 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31568899-a9c0-438f-b273-c54f9ab55ffb-proxy-ca-bundles\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.207253 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31568899-a9c0-438f-b273-c54f9ab55ffb-serving-cert\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.219836 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5tk4\" (UniqueName: \"kubernetes.io/projected/31568899-a9c0-438f-b273-c54f9ab55ffb-kube-api-access-f5tk4\") pod \"controller-manager-7d76f496d8-xcjb6\" (UID: \"31568899-a9c0-438f-b273-c54f9ab55ffb\") " pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.337094 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.575848 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d76f496d8-xcjb6"] Dec 03 17:01:09 crc kubenswrapper[4758]: W1203 17:01:09.588154 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31568899_a9c0_438f_b273_c54f9ab55ffb.slice/crio-f1b11ee1b62270fc4a769752012d1a17902adba537c80e525060d907c7ffefb6 WatchSource:0}: Error finding container f1b11ee1b62270fc4a769752012d1a17902adba537c80e525060d907c7ffefb6: Status 404 returned error can't find the container with id f1b11ee1b62270fc4a769752012d1a17902adba537c80e525060d907c7ffefb6 Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.655380 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" event={"ID":"31568899-a9c0-438f-b273-c54f9ab55ffb","Type":"ContainerStarted","Data":"f1b11ee1b62270fc4a769752012d1a17902adba537c80e525060d907c7ffefb6"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.658549 4758 generic.go:334] "Generic (PLEG): container finished" podID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerID="5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1" exitCode=0 Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.658626 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8ltr" event={"ID":"45bb5bc6-b9e2-403f-9811-8223b2b660cb","Type":"ContainerDied","Data":"5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.664324 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a247927-7fef-4538-b8bb-d3b2d146e318" containerID="78ca1842215e65306acd6413fd7e2ebab6f7b6ed0fd6e9226d12a4375faedbc7" exitCode=0 Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.664399 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26n6q" event={"ID":"3a247927-7fef-4538-b8bb-d3b2d146e318","Type":"ContainerDied","Data":"78ca1842215e65306acd6413fd7e2ebab6f7b6ed0fd6e9226d12a4375faedbc7"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.664427 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26n6q" event={"ID":"3a247927-7fef-4538-b8bb-d3b2d146e318","Type":"ContainerStarted","Data":"253212a9cf20353a123d5829be6c30ebd734d5b768158ddbc3d96f685adac57c"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.667854 4758 generic.go:334] "Generic (PLEG): container finished" podID="dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f" containerID="623c02c0d491f94b943af93a0beae2ac88dfb45df627e0ae50e17bc8b3aab9f2" exitCode=0 Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.667914 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-spff6" event={"ID":"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f","Type":"ContainerDied","Data":"623c02c0d491f94b943af93a0beae2ac88dfb45df627e0ae50e17bc8b3aab9f2"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.667942 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-spff6" event={"ID":"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f","Type":"ContainerStarted","Data":"3084e1f6117a01086e0a551b3fa6fa82eaf37b3709e4b672ec51697a1104548c"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.671429 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t52js" event={"ID":"6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6","Type":"ContainerStarted","Data":"91a7a6adad6de7844ab48a6cc9fbecc746e30828d26c90c6a5cf9a4439f86ecf"} Dec 03 17:01:09 crc kubenswrapper[4758]: I1203 17:01:09.752848 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t52js" podStartSLOduration=2.235422281 podStartE2EDuration="4.752816213s" podCreationTimestamp="2025-12-03 17:01:05 +0000 UTC" firstStartedPulling="2025-12-03 17:01:06.617007697 +0000 UTC m=+321.818384568" lastFinishedPulling="2025-12-03 17:01:09.134401639 +0000 UTC m=+324.335778500" observedRunningTime="2025-12-03 17:01:09.751911938 +0000 UTC m=+324.953288799" watchObservedRunningTime="2025-12-03 17:01:09.752816213 +0000 UTC m=+324.954193074" Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.679385 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" event={"ID":"31568899-a9c0-438f-b273-c54f9ab55ffb","Type":"ContainerStarted","Data":"76f4aa0052e2a7e3c578b2d27daee3684e64efba71e2d27cb5d363f5372ece99"} Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.679854 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.682318 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8ltr" event={"ID":"45bb5bc6-b9e2-403f-9811-8223b2b660cb","Type":"ContainerStarted","Data":"7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb"} Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.684985 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26n6q" event={"ID":"3a247927-7fef-4538-b8bb-d3b2d146e318","Type":"ContainerStarted","Data":"c5bfbaff74b64dab4a9383c81853b0eb7778d071ed9b8002fa0730e765d254d7"} Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.685105 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.688065 4758 generic.go:334] "Generic (PLEG): container finished" podID="dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f" containerID="d421745f6fbe799db0252a47456744d2072d9987ad52d38504ebf0c424180c58" exitCode=0 Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.688115 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-spff6" event={"ID":"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f","Type":"ContainerDied","Data":"d421745f6fbe799db0252a47456744d2072d9987ad52d38504ebf0c424180c58"} Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.738539 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d76f496d8-xcjb6" podStartSLOduration=3.738523294 podStartE2EDuration="3.738523294s" podCreationTimestamp="2025-12-03 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:01:10.712361145 +0000 UTC m=+325.913738016" watchObservedRunningTime="2025-12-03 17:01:10.738523294 +0000 UTC m=+325.939900155" Dec 03 17:01:10 crc kubenswrapper[4758]: I1203 17:01:10.759947 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m8ltr" podStartSLOduration=3.223705972 podStartE2EDuration="5.759923074s" podCreationTimestamp="2025-12-03 17:01:05 +0000 UTC" firstStartedPulling="2025-12-03 17:01:07.625815294 +0000 UTC m=+322.827192165" lastFinishedPulling="2025-12-03 17:01:10.162032406 +0000 UTC m=+325.363409267" observedRunningTime="2025-12-03 17:01:10.756402638 +0000 UTC m=+325.957779489" watchObservedRunningTime="2025-12-03 17:01:10.759923074 +0000 UTC m=+325.961299935" Dec 03 17:01:11 crc kubenswrapper[4758]: I1203 17:01:11.709671 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a247927-7fef-4538-b8bb-d3b2d146e318" containerID="c5bfbaff74b64dab4a9383c81853b0eb7778d071ed9b8002fa0730e765d254d7" exitCode=0 Dec 03 17:01:11 crc kubenswrapper[4758]: I1203 17:01:11.709809 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26n6q" event={"ID":"3a247927-7fef-4538-b8bb-d3b2d146e318","Type":"ContainerDied","Data":"c5bfbaff74b64dab4a9383c81853b0eb7778d071ed9b8002fa0730e765d254d7"} Dec 03 17:01:11 crc kubenswrapper[4758]: I1203 17:01:11.720135 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-spff6" event={"ID":"dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f","Type":"ContainerStarted","Data":"2c221450403ef82241e6e69a6fee16e74431d00bf9cad510b28e6d2105df488f"} Dec 03 17:01:11 crc kubenswrapper[4758]: I1203 17:01:11.751287 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-spff6" podStartSLOduration=3.330682926 podStartE2EDuration="4.75123258s" podCreationTimestamp="2025-12-03 17:01:07 +0000 UTC" firstStartedPulling="2025-12-03 17:01:09.66896104 +0000 UTC m=+324.870337901" lastFinishedPulling="2025-12-03 17:01:11.089510684 +0000 UTC m=+326.290887555" observedRunningTime="2025-12-03 17:01:11.74765213 +0000 UTC m=+326.949029001" watchObservedRunningTime="2025-12-03 17:01:11.75123258 +0000 UTC m=+326.952609461" Dec 03 17:01:12 crc kubenswrapper[4758]: I1203 17:01:12.740850 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26n6q" event={"ID":"3a247927-7fef-4538-b8bb-d3b2d146e318","Type":"ContainerStarted","Data":"d409a57ae997aacee1a5cf1718e338564c4c3dc8f9325461f1839573cda72f70"} Dec 03 17:01:12 crc kubenswrapper[4758]: I1203 17:01:12.768435 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-26n6q" podStartSLOduration=2.258432802 podStartE2EDuration="4.768419301s" podCreationTimestamp="2025-12-03 17:01:08 +0000 UTC" firstStartedPulling="2025-12-03 17:01:09.666576455 +0000 UTC m=+324.867953336" lastFinishedPulling="2025-12-03 17:01:12.176562974 +0000 UTC m=+327.377939835" observedRunningTime="2025-12-03 17:01:12.764928215 +0000 UTC m=+327.966305076" watchObservedRunningTime="2025-12-03 17:01:12.768419301 +0000 UTC m=+327.969796162" Dec 03 17:01:15 crc kubenswrapper[4758]: I1203 17:01:15.916172 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:15 crc kubenswrapper[4758]: I1203 17:01:15.916749 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:15 crc kubenswrapper[4758]: I1203 17:01:15.973844 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:16 crc kubenswrapper[4758]: I1203 17:01:16.095764 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:16 crc kubenswrapper[4758]: I1203 17:01:16.095826 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:16 crc kubenswrapper[4758]: I1203 17:01:16.142703 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:16 crc kubenswrapper[4758]: I1203 17:01:16.812499 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t52js" Dec 03 17:01:16 crc kubenswrapper[4758]: I1203 17:01:16.813803 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.485278 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.486154 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.533527 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.614559 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.615224 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.656698 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.832185 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-spff6" Dec 03 17:01:18 crc kubenswrapper[4758]: I1203 17:01:18.860869 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-26n6q" Dec 03 17:01:26 crc kubenswrapper[4758]: I1203 17:01:26.213993 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bdk6p" Dec 03 17:01:26 crc kubenswrapper[4758]: I1203 17:01:26.284935 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94vfh"] Dec 03 17:01:27 crc kubenswrapper[4758]: I1203 17:01:27.565414 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r"] Dec 03 17:01:27 crc kubenswrapper[4758]: I1203 17:01:27.565645 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" podUID="13c7f556-8630-4320-a7fc-9d7a16f06bb8" containerName="route-controller-manager" containerID="cri-o://5839ac89eab3962554481ddafdfa4ea8510f1d7d7328afa186422a3c63c97883" gracePeriod=30 Dec 03 17:01:27 crc kubenswrapper[4758]: I1203 17:01:27.851856 4758 generic.go:334] "Generic (PLEG): container finished" podID="13c7f556-8630-4320-a7fc-9d7a16f06bb8" containerID="5839ac89eab3962554481ddafdfa4ea8510f1d7d7328afa186422a3c63c97883" exitCode=0 Dec 03 17:01:27 crc kubenswrapper[4758]: I1203 17:01:27.851919 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" event={"ID":"13c7f556-8630-4320-a7fc-9d7a16f06bb8","Type":"ContainerDied","Data":"5839ac89eab3962554481ddafdfa4ea8510f1d7d7328afa186422a3c63c97883"} Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.066082 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.109661 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c7f556-8630-4320-a7fc-9d7a16f06bb8-serving-cert\") pod \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.109788 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8969\" (UniqueName: \"kubernetes.io/projected/13c7f556-8630-4320-a7fc-9d7a16f06bb8-kube-api-access-k8969\") pod \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.109895 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-config\") pod \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.110023 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-client-ca\") pod \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\" (UID: \"13c7f556-8630-4320-a7fc-9d7a16f06bb8\") " Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.110787 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-client-ca" (OuterVolumeSpecName: "client-ca") pod "13c7f556-8630-4320-a7fc-9d7a16f06bb8" (UID: "13c7f556-8630-4320-a7fc-9d7a16f06bb8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.110867 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-config" (OuterVolumeSpecName: "config") pod "13c7f556-8630-4320-a7fc-9d7a16f06bb8" (UID: "13c7f556-8630-4320-a7fc-9d7a16f06bb8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.117075 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c7f556-8630-4320-a7fc-9d7a16f06bb8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "13c7f556-8630-4320-a7fc-9d7a16f06bb8" (UID: "13c7f556-8630-4320-a7fc-9d7a16f06bb8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.117185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c7f556-8630-4320-a7fc-9d7a16f06bb8-kube-api-access-k8969" (OuterVolumeSpecName: "kube-api-access-k8969") pod "13c7f556-8630-4320-a7fc-9d7a16f06bb8" (UID: "13c7f556-8630-4320-a7fc-9d7a16f06bb8"). InnerVolumeSpecName "kube-api-access-k8969". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.211378 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.211439 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c7f556-8630-4320-a7fc-9d7a16f06bb8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.211451 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8969\" (UniqueName: \"kubernetes.io/projected/13c7f556-8630-4320-a7fc-9d7a16f06bb8-kube-api-access-k8969\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.211464 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7f556-8630-4320-a7fc-9d7a16f06bb8-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.861363 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" event={"ID":"13c7f556-8630-4320-a7fc-9d7a16f06bb8","Type":"ContainerDied","Data":"29223a90819738d8c7335c2c5d2088b3e9e9fcfe90d8aaadb3e08685ffadbaa4"} Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.861407 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.861422 4758 scope.go:117] "RemoveContainer" containerID="5839ac89eab3962554481ddafdfa4ea8510f1d7d7328afa186422a3c63c97883" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.896790 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r"] Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.902184 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7f4579db-pp55r"] Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.995554 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm"] Dec 03 17:01:28 crc kubenswrapper[4758]: E1203 17:01:28.996003 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c7f556-8630-4320-a7fc-9d7a16f06bb8" containerName="route-controller-manager" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.996052 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c7f556-8630-4320-a7fc-9d7a16f06bb8" containerName="route-controller-manager" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.996186 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c7f556-8630-4320-a7fc-9d7a16f06bb8" containerName="route-controller-manager" Dec 03 17:01:28 crc kubenswrapper[4758]: I1203 17:01:28.996883 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.003313 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm"] Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.007761 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.008001 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.008034 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.008288 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.008410 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.009535 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.123529 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c7f556-8630-4320-a7fc-9d7a16f06bb8" path="/var/lib/kubelet/pods/13c7f556-8630-4320-a7fc-9d7a16f06bb8/volumes" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.130939 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vztz6\" (UniqueName: \"kubernetes.io/projected/309dd126-b37a-4338-8de4-5c10b35e372f-kube-api-access-vztz6\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.130998 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/309dd126-b37a-4338-8de4-5c10b35e372f-serving-cert\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.131097 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/309dd126-b37a-4338-8de4-5c10b35e372f-client-ca\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.131118 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/309dd126-b37a-4338-8de4-5c10b35e372f-config\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.232513 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/309dd126-b37a-4338-8de4-5c10b35e372f-config\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.233125 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vztz6\" (UniqueName: \"kubernetes.io/projected/309dd126-b37a-4338-8de4-5c10b35e372f-kube-api-access-vztz6\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.234479 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/309dd126-b37a-4338-8de4-5c10b35e372f-serving-cert\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.234703 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/309dd126-b37a-4338-8de4-5c10b35e372f-client-ca\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.235892 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/309dd126-b37a-4338-8de4-5c10b35e372f-client-ca\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.239129 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/309dd126-b37a-4338-8de4-5c10b35e372f-config\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.240861 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/309dd126-b37a-4338-8de4-5c10b35e372f-serving-cert\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.265973 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vztz6\" (UniqueName: \"kubernetes.io/projected/309dd126-b37a-4338-8de4-5c10b35e372f-kube-api-access-vztz6\") pod \"route-controller-manager-5db644644d-mb4wm\" (UID: \"309dd126-b37a-4338-8de4-5c10b35e372f\") " pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.317075 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.779192 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm"] Dec 03 17:01:29 crc kubenswrapper[4758]: I1203 17:01:29.875593 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" event={"ID":"309dd126-b37a-4338-8de4-5c10b35e372f","Type":"ContainerStarted","Data":"44a2a340a3253a99707b3b7d198e3ea4fc2d73b79d28c8c4e7515e1988440b03"} Dec 03 17:01:30 crc kubenswrapper[4758]: I1203 17:01:30.894627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" event={"ID":"309dd126-b37a-4338-8de4-5c10b35e372f","Type":"ContainerStarted","Data":"1f00d3080f0a8bdb628c9c436a398b4dae3bca139fe1c628d5d342af4efc053e"} Dec 03 17:01:30 crc kubenswrapper[4758]: I1203 17:01:30.895087 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:30 crc kubenswrapper[4758]: I1203 17:01:30.901217 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" Dec 03 17:01:30 crc kubenswrapper[4758]: I1203 17:01:30.917618 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5db644644d-mb4wm" podStartSLOduration=3.91759543 podStartE2EDuration="3.91759543s" podCreationTimestamp="2025-12-03 17:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:01:30.914969718 +0000 UTC m=+346.116346579" watchObservedRunningTime="2025-12-03 17:01:30.91759543 +0000 UTC m=+346.118972291" Dec 03 17:01:41 crc kubenswrapper[4758]: I1203 17:01:41.394960 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:01:41 crc kubenswrapper[4758]: I1203 17:01:41.395651 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.332578 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" podUID="992baa6d-67aa-467a-8467-c97d64e69d44" containerName="registry" containerID="cri-o://f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d" gracePeriod=30 Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.817225 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876502 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-bound-sa-token\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876667 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ngkf\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-kube-api-access-7ngkf\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876816 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/992baa6d-67aa-467a-8467-c97d64e69d44-installation-pull-secrets\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876846 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/992baa6d-67aa-467a-8467-c97d64e69d44-ca-trust-extracted\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876870 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-trusted-ca\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876900 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-registry-tls\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.876919 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-registry-certificates\") pod \"992baa6d-67aa-467a-8467-c97d64e69d44\" (UID: \"992baa6d-67aa-467a-8467-c97d64e69d44\") " Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.877837 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.878277 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.883970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-kube-api-access-7ngkf" (OuterVolumeSpecName: "kube-api-access-7ngkf") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "kube-api-access-7ngkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.884183 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.884681 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992baa6d-67aa-467a-8467-c97d64e69d44-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.884971 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.891446 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.895584 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992baa6d-67aa-467a-8467-c97d64e69d44-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "992baa6d-67aa-467a-8467-c97d64e69d44" (UID: "992baa6d-67aa-467a-8467-c97d64e69d44"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.978878 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ngkf\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-kube-api-access-7ngkf\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.978927 4758 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/992baa6d-67aa-467a-8467-c97d64e69d44-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.978947 4758 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/992baa6d-67aa-467a-8467-c97d64e69d44-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.978965 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.978987 4758 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.979003 4758 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/992baa6d-67aa-467a-8467-c97d64e69d44-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:51 crc kubenswrapper[4758]: I1203 17:01:51.979020 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/992baa6d-67aa-467a-8467-c97d64e69d44-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.034450 4758 generic.go:334] "Generic (PLEG): container finished" podID="992baa6d-67aa-467a-8467-c97d64e69d44" containerID="f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d" exitCode=0 Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.034504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" event={"ID":"992baa6d-67aa-467a-8467-c97d64e69d44","Type":"ContainerDied","Data":"f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d"} Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.034566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" event={"ID":"992baa6d-67aa-467a-8467-c97d64e69d44","Type":"ContainerDied","Data":"2452f2c717c8fc2a8a6b8b83b76e6ea8db99e62dc95a929183175c56b03bd176"} Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.034581 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94vfh" Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.034591 4758 scope.go:117] "RemoveContainer" containerID="f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d" Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.062364 4758 scope.go:117] "RemoveContainer" containerID="f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d" Dec 03 17:01:52 crc kubenswrapper[4758]: E1203 17:01:52.063896 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d\": container with ID starting with f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d not found: ID does not exist" containerID="f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d" Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.063956 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d"} err="failed to get container status \"f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d\": rpc error: code = NotFound desc = could not find container \"f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d\": container with ID starting with f7fe3aa4621f3a26510d3f2c5f6531fcf973dbce78914f7c9f69ad67f08dc46d not found: ID does not exist" Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.075136 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94vfh"] Dec 03 17:01:52 crc kubenswrapper[4758]: I1203 17:01:52.080385 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94vfh"] Dec 03 17:01:53 crc kubenswrapper[4758]: I1203 17:01:53.137653 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="992baa6d-67aa-467a-8467-c97d64e69d44" path="/var/lib/kubelet/pods/992baa6d-67aa-467a-8467-c97d64e69d44/volumes" Dec 03 17:02:11 crc kubenswrapper[4758]: I1203 17:02:11.395583 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:02:11 crc kubenswrapper[4758]: I1203 17:02:11.396774 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:02:41 crc kubenswrapper[4758]: I1203 17:02:41.395109 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:02:41 crc kubenswrapper[4758]: I1203 17:02:41.395793 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:02:41 crc kubenswrapper[4758]: I1203 17:02:41.395864 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:02:41 crc kubenswrapper[4758]: I1203 17:02:41.396765 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64765cfa70b0509666919f46ee9d2002e45bb95cfbe37b712a0e15a96f9a4223"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:02:41 crc kubenswrapper[4758]: I1203 17:02:41.396833 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://64765cfa70b0509666919f46ee9d2002e45bb95cfbe37b712a0e15a96f9a4223" gracePeriod=600 Dec 03 17:02:42 crc kubenswrapper[4758]: I1203 17:02:42.369094 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="64765cfa70b0509666919f46ee9d2002e45bb95cfbe37b712a0e15a96f9a4223" exitCode=0 Dec 03 17:02:42 crc kubenswrapper[4758]: I1203 17:02:42.369323 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"64765cfa70b0509666919f46ee9d2002e45bb95cfbe37b712a0e15a96f9a4223"} Dec 03 17:02:42 crc kubenswrapper[4758]: I1203 17:02:42.369426 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"cf08df8af9e9299d90e382d0abba32022a6918755ebe224d4bd41c1289a7b799"} Dec 03 17:02:42 crc kubenswrapper[4758]: I1203 17:02:42.369446 4758 scope.go:117] "RemoveContainer" containerID="f55ff0027dd0f8ea26de7a6d70bb7d37a1dbd60b302ad361798f83301c5b8259" Dec 03 17:05:11 crc kubenswrapper[4758]: I1203 17:05:11.395451 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:05:11 crc kubenswrapper[4758]: I1203 17:05:11.396333 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:05:41 crc kubenswrapper[4758]: I1203 17:05:41.395287 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:05:41 crc kubenswrapper[4758]: I1203 17:05:41.396873 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:06:11 crc kubenswrapper[4758]: I1203 17:06:11.394664 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:06:11 crc kubenswrapper[4758]: I1203 17:06:11.395709 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:06:11 crc kubenswrapper[4758]: I1203 17:06:11.395782 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:06:11 crc kubenswrapper[4758]: I1203 17:06:11.396742 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf08df8af9e9299d90e382d0abba32022a6918755ebe224d4bd41c1289a7b799"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:06:11 crc kubenswrapper[4758]: I1203 17:06:11.396836 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://cf08df8af9e9299d90e382d0abba32022a6918755ebe224d4bd41c1289a7b799" gracePeriod=600 Dec 03 17:06:12 crc kubenswrapper[4758]: I1203 17:06:12.127579 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="cf08df8af9e9299d90e382d0abba32022a6918755ebe224d4bd41c1289a7b799" exitCode=0 Dec 03 17:06:12 crc kubenswrapper[4758]: I1203 17:06:12.128255 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"cf08df8af9e9299d90e382d0abba32022a6918755ebe224d4bd41c1289a7b799"} Dec 03 17:06:12 crc kubenswrapper[4758]: I1203 17:06:12.128418 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"174745d2d4c87b79a75b724a8f3fd0dcd39be458fd52fd0a6f156eafe510a247"} Dec 03 17:06:12 crc kubenswrapper[4758]: I1203 17:06:12.128561 4758 scope.go:117] "RemoveContainer" containerID="64765cfa70b0509666919f46ee9d2002e45bb95cfbe37b712a0e15a96f9a4223" Dec 03 17:08:11 crc kubenswrapper[4758]: I1203 17:08:11.394461 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:08:11 crc kubenswrapper[4758]: I1203 17:08:11.395047 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:08:12 crc kubenswrapper[4758]: I1203 17:08:12.778217 4758 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 17:08:41 crc kubenswrapper[4758]: I1203 17:08:41.395256 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:08:41 crc kubenswrapper[4758]: I1203 17:08:41.396008 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.896189 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9p24"] Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.897308 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-controller" containerID="cri-o://966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.897777 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="sbdb" containerID="cri-o://f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.897834 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="nbdb" containerID="cri-o://6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.897876 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="northd" containerID="cri-o://b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.897925 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.897973 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-node" containerID="cri-o://9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.898019 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-acl-logging" containerID="cri-o://75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05" gracePeriod=30 Dec 03 17:08:57 crc kubenswrapper[4758]: I1203 17:08:57.935458 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" containerID="cri-o://b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958" gracePeriod=30 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.224511 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovnkube-controller/3.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.228492 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovn-acl-logging/0.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.229484 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovn-controller/0.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230186 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958" exitCode=0 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230451 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a" exitCode=0 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230613 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a" exitCode=0 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230748 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2" exitCode=0 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230900 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f" exitCode=0 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230997 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05" exitCode=143 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231069 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1" exitCode=143 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.230271 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231196 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231233 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231252 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231251 4758 scope.go:117] "RemoveContainer" containerID="95a231a9372ab4c857c6fe1963e89b1765eb5f71438df7029b9be41fa973a6de" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231265 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.231451 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.236117 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/2.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.237130 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/1.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.237328 4758 generic.go:334] "Generic (PLEG): container finished" podID="4b5429d1-1fc3-4603-93ba-b57b33c2f585" containerID="01a2e2844e1250fbdb9021acd40f8f5ec395720ebc71734423c63cdf86bde9fd" exitCode=2 Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.237396 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerDied","Data":"01a2e2844e1250fbdb9021acd40f8f5ec395720ebc71734423c63cdf86bde9fd"} Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.238261 4758 scope.go:117] "RemoveContainer" containerID="01a2e2844e1250fbdb9021acd40f8f5ec395720ebc71734423c63cdf86bde9fd" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.269089 4758 scope.go:117] "RemoveContainer" containerID="949f49543a9ec951b7ac389b4b7573a2f55e262bd5f37ba512aa281032092a9c" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.661328 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovn-acl-logging/0.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.662560 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovn-controller/0.log" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.663232 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.720761 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rjblg"] Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721017 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721045 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721064 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721073 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721083 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="northd" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721094 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="northd" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721102 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721110 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721120 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="nbdb" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721130 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="nbdb" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721141 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="sbdb" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721149 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="sbdb" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721163 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992baa6d-67aa-467a-8467-c97d64e69d44" containerName="registry" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721171 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="992baa6d-67aa-467a-8467-c97d64e69d44" containerName="registry" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721183 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kubecfg-setup" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721192 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kubecfg-setup" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721203 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-acl-logging" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721211 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-acl-logging" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721223 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721232 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721243 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-node" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721251 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-node" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721264 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721273 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721408 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721424 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="sbdb" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721434 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721447 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-node" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721458 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="nbdb" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721473 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721483 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721493 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721504 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovn-acl-logging" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721514 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="992baa6d-67aa-467a-8467-c97d64e69d44" containerName="registry" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721527 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="northd" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721538 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721673 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721710 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: E1203 17:08:58.721722 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721733 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.721879 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerName="ovnkube-controller" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.724075 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800534 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-node-log\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-etc-openvswitch\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800592 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-slash\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800626 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47xn2\" (UniqueName: \"kubernetes.io/projected/b40dc3c2-0bc6-447b-ae72-d71650d53f58-kube-api-access-47xn2\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800657 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-script-lib\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800704 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-config\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800727 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-var-lib-openvswitch\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800760 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-ovn\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800699 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-slash" (OuterVolumeSpecName: "host-slash") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800699 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800693 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-node-log" (OuterVolumeSpecName: "node-log") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800774 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800800 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-systemd\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800923 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-openvswitch\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.800953 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801025 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-env-overrides\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801034 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801055 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-bin\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801085 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovn-node-metrics-cert\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801120 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-log-socket\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801114 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801153 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801140 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-systemd-units\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801175 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801164 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801182 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801209 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-log-socket" (OuterVolumeSpecName: "log-socket") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801216 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-ovn-kubernetes\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801241 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-kubelet\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801268 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-netd\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801240 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801293 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-netns\") pod \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\" (UID: \"b40dc3c2-0bc6-447b-ae72-d71650d53f58\") " Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801312 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801334 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801356 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801503 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801652 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801665 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801673 4758 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801723 4758 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801732 4758 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801741 4758 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801751 4758 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b40dc3c2-0bc6-447b-ae72-d71650d53f58-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801760 4758 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801772 4758 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801782 4758 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801793 4758 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801807 4758 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801818 4758 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801831 4758 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801842 4758 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801854 4758 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.801864 4758 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.806533 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.806732 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40dc3c2-0bc6-447b-ae72-d71650d53f58-kube-api-access-47xn2" (OuterVolumeSpecName: "kube-api-access-47xn2") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "kube-api-access-47xn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.813185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b40dc3c2-0bc6-447b-ae72-d71650d53f58" (UID: "b40dc3c2-0bc6-447b-ae72-d71650d53f58"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903349 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903408 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-run-ovn-kubernetes\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903448 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-ovn\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903475 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-node-log\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903536 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-log-socket\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903563 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-cni-netd\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903665 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-kubelet\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903757 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-run-netns\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903784 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btr8v\" (UniqueName: \"kubernetes.io/projected/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-kube-api-access-btr8v\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903810 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovn-node-metrics-cert\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903833 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903851 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-env-overrides\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-systemd-units\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903932 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-systemd\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-var-lib-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.903995 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-etc-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904016 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovnkube-config\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904043 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-cni-bin\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904095 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovnkube-script-lib\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904132 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-slash\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904235 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47xn2\" (UniqueName: \"kubernetes.io/projected/b40dc3c2-0bc6-447b-ae72-d71650d53f58-kube-api-access-47xn2\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904255 4758 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b40dc3c2-0bc6-447b-ae72-d71650d53f58-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:58 crc kubenswrapper[4758]: I1203 17:08:58.904268 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b40dc3c2-0bc6-447b-ae72-d71650d53f58-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005591 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-ovn\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005660 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-node-log\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005723 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-log-socket\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005773 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-cni-netd\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005808 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-kubelet\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005828 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-run-netns\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005851 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btr8v\" (UniqueName: \"kubernetes.io/projected/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-kube-api-access-btr8v\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005875 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovn-node-metrics-cert\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005893 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-env-overrides\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005931 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-systemd-units\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-systemd\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005970 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-var-lib-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.005990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovnkube-config\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006006 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-etc-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006022 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-cni-bin\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006051 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovnkube-script-lib\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006069 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-slash\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006098 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006116 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-run-ovn-kubernetes\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006184 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-run-ovn-kubernetes\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006222 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-ovn\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006242 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-node-log\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006261 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-log-socket\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006281 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-cni-netd\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006302 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-kubelet\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.006320 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-run-netns\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008146 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-systemd\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008219 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-etc-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008293 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-var-lib-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008300 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-slash\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008282 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-run-openvswitch\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008318 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-cni-bin\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008335 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-systemd-units\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.008558 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovnkube-config\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.009156 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-env-overrides\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.009188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovnkube-script-lib\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.011757 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-ovn-node-metrics-cert\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.030051 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btr8v\" (UniqueName: \"kubernetes.io/projected/b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc-kube-api-access-btr8v\") pod \"ovnkube-node-rjblg\" (UID: \"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.041311 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:08:59 crc kubenswrapper[4758]: W1203 17:08:59.061262 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9bd0c9c_be32_45d6_bb5c_01ceb11e13fc.slice/crio-3495fc2f447c0553c52bc17ddd40b89bd384a5576681c62997783c4142e3f68e WatchSource:0}: Error finding container 3495fc2f447c0553c52bc17ddd40b89bd384a5576681c62997783c4142e3f68e: Status 404 returned error can't find the container with id 3495fc2f447c0553c52bc17ddd40b89bd384a5576681c62997783c4142e3f68e Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.248067 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovn-acl-logging/0.log" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.248738 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9p24_b40dc3c2-0bc6-447b-ae72-d71650d53f58/ovn-controller/0.log" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.249229 4758 generic.go:334] "Generic (PLEG): container finished" podID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" containerID="f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d" exitCode=0 Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.249416 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d"} Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.249534 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" event={"ID":"b40dc3c2-0bc6-447b-ae72-d71650d53f58","Type":"ContainerDied","Data":"c3de506de982657a3e05ab139a6d51b0375cba0005e39dbaba3b902b37ea6dce"} Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.249559 4758 scope.go:117] "RemoveContainer" containerID="b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.250118 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9p24" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.251445 4758 generic.go:334] "Generic (PLEG): container finished" podID="b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc" containerID="a3d05243000494b9f8d98550dad378793fd2c849fe8a0bc34219c802c83ef49d" exitCode=0 Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.251512 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerDied","Data":"a3d05243000494b9f8d98550dad378793fd2c849fe8a0bc34219c802c83ef49d"} Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.251529 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"3495fc2f447c0553c52bc17ddd40b89bd384a5576681c62997783c4142e3f68e"} Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.254475 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zwcmr_4b5429d1-1fc3-4603-93ba-b57b33c2f585/kube-multus/2.log" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.254664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zwcmr" event={"ID":"4b5429d1-1fc3-4603-93ba-b57b33c2f585","Type":"ContainerStarted","Data":"ac5ed88d4a8a9ddbcbcfed77b6edce05a742ad17aa02e4231cf78774d04eee1e"} Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.298533 4758 scope.go:117] "RemoveContainer" containerID="f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.337189 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9p24"] Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.342131 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9p24"] Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.345976 4758 scope.go:117] "RemoveContainer" containerID="6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.363143 4758 scope.go:117] "RemoveContainer" containerID="b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.387460 4758 scope.go:117] "RemoveContainer" containerID="393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.406977 4758 scope.go:117] "RemoveContainer" containerID="9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.426861 4758 scope.go:117] "RemoveContainer" containerID="75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.444990 4758 scope.go:117] "RemoveContainer" containerID="966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.466437 4758 scope.go:117] "RemoveContainer" containerID="23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.496429 4758 scope.go:117] "RemoveContainer" containerID="b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.497003 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958\": container with ID starting with b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958 not found: ID does not exist" containerID="b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.497042 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958"} err="failed to get container status \"b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958\": rpc error: code = NotFound desc = could not find container \"b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958\": container with ID starting with b16ce315a5bee197639083ea733272a22ec0267d74664eebed305f71ccbb6958 not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.497071 4758 scope.go:117] "RemoveContainer" containerID="f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.497369 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\": container with ID starting with f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d not found: ID does not exist" containerID="f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.497432 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d"} err="failed to get container status \"f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\": rpc error: code = NotFound desc = could not find container \"f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d\": container with ID starting with f247b9f995f5d13be661d3ba1b8335b31d5c39c87803a4c362ee4c3615baa90d not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.497482 4758 scope.go:117] "RemoveContainer" containerID="6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.497914 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\": container with ID starting with 6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a not found: ID does not exist" containerID="6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.497967 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a"} err="failed to get container status \"6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\": rpc error: code = NotFound desc = could not find container \"6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a\": container with ID starting with 6dba082c3c6d1d2828d50ca3f1337812ed5621c9797d2d825f7abb6ece9f951a not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.498002 4758 scope.go:117] "RemoveContainer" containerID="b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.498365 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\": container with ID starting with b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a not found: ID does not exist" containerID="b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.498487 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a"} err="failed to get container status \"b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\": rpc error: code = NotFound desc = could not find container \"b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a\": container with ID starting with b404704f9e8768f9fd05f9e00c095940bfe0743e1b3210323ff5b140ef116c6a not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.498513 4758 scope.go:117] "RemoveContainer" containerID="393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.498776 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\": container with ID starting with 393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2 not found: ID does not exist" containerID="393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.498817 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2"} err="failed to get container status \"393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\": rpc error: code = NotFound desc = could not find container \"393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2\": container with ID starting with 393d40e0c8767c5c4b757b0026e9c81e954bff4ffd63c06043eca47bf3d89dc2 not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.498844 4758 scope.go:117] "RemoveContainer" containerID="9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.499116 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\": container with ID starting with 9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f not found: ID does not exist" containerID="9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.499209 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f"} err="failed to get container status \"9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\": rpc error: code = NotFound desc = could not find container \"9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f\": container with ID starting with 9d6bcbd3d33c59efe38263676115f30784c695ade0e192062a09110e312afe0f not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.499267 4758 scope.go:117] "RemoveContainer" containerID="75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.499563 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\": container with ID starting with 75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05 not found: ID does not exist" containerID="75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.499606 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05"} err="failed to get container status \"75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\": rpc error: code = NotFound desc = could not find container \"75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05\": container with ID starting with 75fc5e87ef16ef4989e7a09ba6472d95a76c4b61bbb7fc911418d61c8f8b6c05 not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.499633 4758 scope.go:117] "RemoveContainer" containerID="966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.499903 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\": container with ID starting with 966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1 not found: ID does not exist" containerID="966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.499934 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1"} err="failed to get container status \"966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\": rpc error: code = NotFound desc = could not find container \"966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1\": container with ID starting with 966a5018a5226009b2b777621458f18375193577ac87a2a5366bc340d34cd9c1 not found: ID does not exist" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.499953 4758 scope.go:117] "RemoveContainer" containerID="23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3" Dec 03 17:08:59 crc kubenswrapper[4758]: E1203 17:08:59.500167 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\": container with ID starting with 23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3 not found: ID does not exist" containerID="23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3" Dec 03 17:08:59 crc kubenswrapper[4758]: I1203 17:08:59.500195 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3"} err="failed to get container status \"23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\": rpc error: code = NotFound desc = could not find container \"23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3\": container with ID starting with 23737cd19338115e192dc29e8d200ee1feb9a39a5cdc8817a917e434019d39b3 not found: ID does not exist" Dec 03 17:09:00 crc kubenswrapper[4758]: I1203 17:09:00.264712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"8624fe1cf05c4e54de440795a852168fff439105667b83aa4284c34e8aea8617"} Dec 03 17:09:00 crc kubenswrapper[4758]: I1203 17:09:00.266568 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"3b8093a782d7241cbce23c18dbec9b687098ff41f8ffb5713cb9b86c237d7831"} Dec 03 17:09:00 crc kubenswrapper[4758]: I1203 17:09:00.266666 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"2b440b5f564c1f3215c66894bb8bb79956cb4f197fd14ec3348901eb5eba5ed9"} Dec 03 17:09:00 crc kubenswrapper[4758]: I1203 17:09:00.266764 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"519b4ef022185d237c7ed44978c858b0352ded75bfde860a6571f86776aade0f"} Dec 03 17:09:00 crc kubenswrapper[4758]: I1203 17:09:00.266822 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"107d5c5a148b481d608e3df62185961689f21039b7934f3379573b779f4adb22"} Dec 03 17:09:00 crc kubenswrapper[4758]: I1203 17:09:00.266882 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"52d0dc70cf725df5bacc75f4ef04b76025cc7486061d0b7e33f8399edb5a762b"} Dec 03 17:09:01 crc kubenswrapper[4758]: I1203 17:09:01.126358 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40dc3c2-0bc6-447b-ae72-d71650d53f58" path="/var/lib/kubelet/pods/b40dc3c2-0bc6-447b-ae72-d71650d53f58/volumes" Dec 03 17:09:03 crc kubenswrapper[4758]: I1203 17:09:03.286924 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"2b0a05d214b6e4f49fab9e9604a88efa3f9358629763c597f705dc7f9bc954b5"} Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.068585 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-mhpzq"] Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.070105 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.072673 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.072973 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.073512 4758 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-ggqk7" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.073543 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.218061 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/85230d29-9148-4067-a103-2482c48013fb-node-mnt\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.218165 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/85230d29-9148-4067-a103-2482c48013fb-crc-storage\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.218194 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm2tl\" (UniqueName: \"kubernetes.io/projected/85230d29-9148-4067-a103-2482c48013fb-kube-api-access-pm2tl\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.320078 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/85230d29-9148-4067-a103-2482c48013fb-node-mnt\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.320189 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/85230d29-9148-4067-a103-2482c48013fb-crc-storage\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.320223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm2tl\" (UniqueName: \"kubernetes.io/projected/85230d29-9148-4067-a103-2482c48013fb-kube-api-access-pm2tl\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.320500 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/85230d29-9148-4067-a103-2482c48013fb-node-mnt\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.321483 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/85230d29-9148-4067-a103-2482c48013fb-crc-storage\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.349307 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm2tl\" (UniqueName: \"kubernetes.io/projected/85230d29-9148-4067-a103-2482c48013fb-kube-api-access-pm2tl\") pod \"crc-storage-crc-mhpzq\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: I1203 17:09:07.386274 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: E1203 17:09:07.418402 4758 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(94888ff8c563a064f0a2b576a1a14b35b7de8c23092f5ca0d884aaea391a0d73): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:09:07 crc kubenswrapper[4758]: E1203 17:09:07.418539 4758 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(94888ff8c563a064f0a2b576a1a14b35b7de8c23092f5ca0d884aaea391a0d73): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: E1203 17:09:07.418585 4758 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(94888ff8c563a064f0a2b576a1a14b35b7de8c23092f5ca0d884aaea391a0d73): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:07 crc kubenswrapper[4758]: E1203 17:09:07.418657 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-mhpzq_crc-storage(85230d29-9148-4067-a103-2482c48013fb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-mhpzq_crc-storage(85230d29-9148-4067-a103-2482c48013fb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(94888ff8c563a064f0a2b576a1a14b35b7de8c23092f5ca0d884aaea391a0d73): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-mhpzq" podUID="85230d29-9148-4067-a103-2482c48013fb" Dec 03 17:09:08 crc kubenswrapper[4758]: I1203 17:09:08.325991 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" event={"ID":"b9bd0c9c-be32-45d6-bb5c-01ceb11e13fc","Type":"ContainerStarted","Data":"7f4a2a75153f59d984bb1994e994212bcccdad9e813b4991ece3d4f1e4c2a4e9"} Dec 03 17:09:08 crc kubenswrapper[4758]: I1203 17:09:08.382086 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-mhpzq"] Dec 03 17:09:08 crc kubenswrapper[4758]: I1203 17:09:08.382187 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:08 crc kubenswrapper[4758]: I1203 17:09:08.382671 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:08 crc kubenswrapper[4758]: E1203 17:09:08.408721 4758 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(a88ed0ac6b8a5d711ed20b08aaac47bf8384342c0a0ccd8ebd4254a96dc754dd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:09:08 crc kubenswrapper[4758]: E1203 17:09:08.408795 4758 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(a88ed0ac6b8a5d711ed20b08aaac47bf8384342c0a0ccd8ebd4254a96dc754dd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:08 crc kubenswrapper[4758]: E1203 17:09:08.408827 4758 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(a88ed0ac6b8a5d711ed20b08aaac47bf8384342c0a0ccd8ebd4254a96dc754dd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:08 crc kubenswrapper[4758]: E1203 17:09:08.408875 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-mhpzq_crc-storage(85230d29-9148-4067-a103-2482c48013fb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-mhpzq_crc-storage(85230d29-9148-4067-a103-2482c48013fb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-mhpzq_crc-storage_85230d29-9148-4067-a103-2482c48013fb_0(a88ed0ac6b8a5d711ed20b08aaac47bf8384342c0a0ccd8ebd4254a96dc754dd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-mhpzq" podUID="85230d29-9148-4067-a103-2482c48013fb" Dec 03 17:09:09 crc kubenswrapper[4758]: I1203 17:09:09.332416 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:09:09 crc kubenswrapper[4758]: I1203 17:09:09.333798 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:09:09 crc kubenswrapper[4758]: I1203 17:09:09.333813 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:09:09 crc kubenswrapper[4758]: I1203 17:09:09.371297 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:09:09 crc kubenswrapper[4758]: I1203 17:09:09.371878 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:09:09 crc kubenswrapper[4758]: I1203 17:09:09.375369 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" podStartSLOduration=11.37534016 podStartE2EDuration="11.37534016s" podCreationTimestamp="2025-12-03 17:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:09:09.370960364 +0000 UTC m=+804.572337225" watchObservedRunningTime="2025-12-03 17:09:09.37534016 +0000 UTC m=+804.576717021" Dec 03 17:09:11 crc kubenswrapper[4758]: I1203 17:09:11.394614 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:09:11 crc kubenswrapper[4758]: I1203 17:09:11.395332 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:09:11 crc kubenswrapper[4758]: I1203 17:09:11.395409 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:09:11 crc kubenswrapper[4758]: I1203 17:09:11.396582 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"174745d2d4c87b79a75b724a8f3fd0dcd39be458fd52fd0a6f156eafe510a247"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:09:11 crc kubenswrapper[4758]: I1203 17:09:11.396706 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://174745d2d4c87b79a75b724a8f3fd0dcd39be458fd52fd0a6f156eafe510a247" gracePeriod=600 Dec 03 17:09:12 crc kubenswrapper[4758]: I1203 17:09:12.368571 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="174745d2d4c87b79a75b724a8f3fd0dcd39be458fd52fd0a6f156eafe510a247" exitCode=0 Dec 03 17:09:12 crc kubenswrapper[4758]: I1203 17:09:12.368663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"174745d2d4c87b79a75b724a8f3fd0dcd39be458fd52fd0a6f156eafe510a247"} Dec 03 17:09:12 crc kubenswrapper[4758]: I1203 17:09:12.369231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"8050d5b5272390ff909cae77aca4e99c2b09448e05fbe4bc6797c541002e26b0"} Dec 03 17:09:12 crc kubenswrapper[4758]: I1203 17:09:12.369281 4758 scope.go:117] "RemoveContainer" containerID="cf08df8af9e9299d90e382d0abba32022a6918755ebe224d4bd41c1289a7b799" Dec 03 17:09:23 crc kubenswrapper[4758]: I1203 17:09:23.114335 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:23 crc kubenswrapper[4758]: I1203 17:09:23.115428 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:23 crc kubenswrapper[4758]: I1203 17:09:23.308373 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-mhpzq"] Dec 03 17:09:23 crc kubenswrapper[4758]: I1203 17:09:23.322551 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:09:23 crc kubenswrapper[4758]: I1203 17:09:23.457843 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-mhpzq" event={"ID":"85230d29-9148-4067-a103-2482c48013fb","Type":"ContainerStarted","Data":"cd4f8d51ecf99a4e2c869cf82ce776d645a3f440d5312cbd9bcebd4f476cf298"} Dec 03 17:09:25 crc kubenswrapper[4758]: I1203 17:09:25.474878 4758 generic.go:334] "Generic (PLEG): container finished" podID="85230d29-9148-4067-a103-2482c48013fb" containerID="9b4c1858e6eb9e695919edccf921e60150fbafa35f743e2d599dee47e401775a" exitCode=0 Dec 03 17:09:25 crc kubenswrapper[4758]: I1203 17:09:25.475032 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-mhpzq" event={"ID":"85230d29-9148-4067-a103-2482c48013fb","Type":"ContainerDied","Data":"9b4c1858e6eb9e695919edccf921e60150fbafa35f743e2d599dee47e401775a"} Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.733041 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.765413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm2tl\" (UniqueName: \"kubernetes.io/projected/85230d29-9148-4067-a103-2482c48013fb-kube-api-access-pm2tl\") pod \"85230d29-9148-4067-a103-2482c48013fb\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.765527 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/85230d29-9148-4067-a103-2482c48013fb-crc-storage\") pod \"85230d29-9148-4067-a103-2482c48013fb\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.765579 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/85230d29-9148-4067-a103-2482c48013fb-node-mnt\") pod \"85230d29-9148-4067-a103-2482c48013fb\" (UID: \"85230d29-9148-4067-a103-2482c48013fb\") " Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.765871 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/85230d29-9148-4067-a103-2482c48013fb-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "85230d29-9148-4067-a103-2482c48013fb" (UID: "85230d29-9148-4067-a103-2482c48013fb"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.772625 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85230d29-9148-4067-a103-2482c48013fb-kube-api-access-pm2tl" (OuterVolumeSpecName: "kube-api-access-pm2tl") pod "85230d29-9148-4067-a103-2482c48013fb" (UID: "85230d29-9148-4067-a103-2482c48013fb"). InnerVolumeSpecName "kube-api-access-pm2tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.783223 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85230d29-9148-4067-a103-2482c48013fb-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "85230d29-9148-4067-a103-2482c48013fb" (UID: "85230d29-9148-4067-a103-2482c48013fb"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.866572 4758 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/85230d29-9148-4067-a103-2482c48013fb-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.866640 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm2tl\" (UniqueName: \"kubernetes.io/projected/85230d29-9148-4067-a103-2482c48013fb-kube-api-access-pm2tl\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:26 crc kubenswrapper[4758]: I1203 17:09:26.866654 4758 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/85230d29-9148-4067-a103-2482c48013fb-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:27 crc kubenswrapper[4758]: I1203 17:09:27.490470 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-mhpzq" event={"ID":"85230d29-9148-4067-a103-2482c48013fb","Type":"ContainerDied","Data":"cd4f8d51ecf99a4e2c869cf82ce776d645a3f440d5312cbd9bcebd4f476cf298"} Dec 03 17:09:27 crc kubenswrapper[4758]: I1203 17:09:27.490542 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd4f8d51ecf99a4e2c869cf82ce776d645a3f440d5312cbd9bcebd4f476cf298" Dec 03 17:09:27 crc kubenswrapper[4758]: I1203 17:09:27.491253 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mhpzq" Dec 03 17:09:29 crc kubenswrapper[4758]: I1203 17:09:29.072714 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rjblg" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.083173 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2"] Dec 03 17:09:35 crc kubenswrapper[4758]: E1203 17:09:35.084198 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85230d29-9148-4067-a103-2482c48013fb" containerName="storage" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.084213 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="85230d29-9148-4067-a103-2482c48013fb" containerName="storage" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.084336 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="85230d29-9148-4067-a103-2482c48013fb" containerName="storage" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.085150 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.088034 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.101849 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2"] Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.182447 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz9j7\" (UniqueName: \"kubernetes.io/projected/ad1568bc-9c16-4426-b740-c6ac65af84c0-kube-api-access-cz9j7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.182547 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.182726 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.283854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.283945 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz9j7\" (UniqueName: \"kubernetes.io/projected/ad1568bc-9c16-4426-b740-c6ac65af84c0-kube-api-access-cz9j7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.283993 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.284666 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.284723 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.306068 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz9j7\" (UniqueName: \"kubernetes.io/projected/ad1568bc-9c16-4426-b740-c6ac65af84c0-kube-api-access-cz9j7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.404260 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:35 crc kubenswrapper[4758]: I1203 17:09:35.607733 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2"] Dec 03 17:09:36 crc kubenswrapper[4758]: I1203 17:09:36.549485 4758 generic.go:334] "Generic (PLEG): container finished" podID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerID="b3e59e949405394c87a009cd22d663487c15dd0118c5f072d40428f0dce05fe9" exitCode=0 Dec 03 17:09:36 crc kubenswrapper[4758]: I1203 17:09:36.549608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" event={"ID":"ad1568bc-9c16-4426-b740-c6ac65af84c0","Type":"ContainerDied","Data":"b3e59e949405394c87a009cd22d663487c15dd0118c5f072d40428f0dce05fe9"} Dec 03 17:09:36 crc kubenswrapper[4758]: I1203 17:09:36.550205 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" event={"ID":"ad1568bc-9c16-4426-b740-c6ac65af84c0","Type":"ContainerStarted","Data":"b81791c3945f9ff7fb27793de8ab3534fe7e79bcb33037eff32a1315f2b12c79"} Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.082814 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nxffg"] Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.084013 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.097650 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxffg"] Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.211870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcrlc\" (UniqueName: \"kubernetes.io/projected/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-kube-api-access-fcrlc\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.212935 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-catalog-content\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.212969 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-utilities\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.314882 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-catalog-content\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.314943 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-utilities\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.315031 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcrlc\" (UniqueName: \"kubernetes.io/projected/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-kube-api-access-fcrlc\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.315417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-catalog-content\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.315497 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-utilities\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.339843 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcrlc\" (UniqueName: \"kubernetes.io/projected/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-kube-api-access-fcrlc\") pod \"redhat-operators-nxffg\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.419898 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:37 crc kubenswrapper[4758]: I1203 17:09:37.639871 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxffg"] Dec 03 17:09:37 crc kubenswrapper[4758]: W1203 17:09:37.647156 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa036f9c_f46c_4045_9f0b_7c880b62f1f8.slice/crio-e9bb0cdcd72cd82c1c80c35481a9d5285a102d6029ca50520e3bcd829548e188 WatchSource:0}: Error finding container e9bb0cdcd72cd82c1c80c35481a9d5285a102d6029ca50520e3bcd829548e188: Status 404 returned error can't find the container with id e9bb0cdcd72cd82c1c80c35481a9d5285a102d6029ca50520e3bcd829548e188 Dec 03 17:09:38 crc kubenswrapper[4758]: I1203 17:09:38.567205 4758 generic.go:334] "Generic (PLEG): container finished" podID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerID="fbe91367b18a2565736167f76c56489c069ad3c51e1c36746d92bc7b0ff322ca" exitCode=0 Dec 03 17:09:38 crc kubenswrapper[4758]: I1203 17:09:38.567279 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerDied","Data":"fbe91367b18a2565736167f76c56489c069ad3c51e1c36746d92bc7b0ff322ca"} Dec 03 17:09:38 crc kubenswrapper[4758]: I1203 17:09:38.567638 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerStarted","Data":"e9bb0cdcd72cd82c1c80c35481a9d5285a102d6029ca50520e3bcd829548e188"} Dec 03 17:09:39 crc kubenswrapper[4758]: I1203 17:09:39.576664 4758 generic.go:334] "Generic (PLEG): container finished" podID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerID="8e5baf26fbe87b6358ef2626288ee7680b670f5b24ae4fbddf33912d58b2ae96" exitCode=0 Dec 03 17:09:39 crc kubenswrapper[4758]: I1203 17:09:39.576858 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" event={"ID":"ad1568bc-9c16-4426-b740-c6ac65af84c0","Type":"ContainerDied","Data":"8e5baf26fbe87b6358ef2626288ee7680b670f5b24ae4fbddf33912d58b2ae96"} Dec 03 17:09:39 crc kubenswrapper[4758]: I1203 17:09:39.580142 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerStarted","Data":"b04b5c1bdc46c6b766455c52a1c1439cb612a1b2bde98a3a2a3d590c5d8da9d3"} Dec 03 17:09:40 crc kubenswrapper[4758]: I1203 17:09:40.586144 4758 generic.go:334] "Generic (PLEG): container finished" podID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerID="b04b5c1bdc46c6b766455c52a1c1439cb612a1b2bde98a3a2a3d590c5d8da9d3" exitCode=0 Dec 03 17:09:40 crc kubenswrapper[4758]: I1203 17:09:40.586211 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerDied","Data":"b04b5c1bdc46c6b766455c52a1c1439cb612a1b2bde98a3a2a3d590c5d8da9d3"} Dec 03 17:09:40 crc kubenswrapper[4758]: I1203 17:09:40.591403 4758 generic.go:334] "Generic (PLEG): container finished" podID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerID="495c6fecbb43e66901d319970bdee31e76b4585ad7e926173f59dbf42a185adb" exitCode=0 Dec 03 17:09:40 crc kubenswrapper[4758]: I1203 17:09:40.591485 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" event={"ID":"ad1568bc-9c16-4426-b740-c6ac65af84c0","Type":"ContainerDied","Data":"495c6fecbb43e66901d319970bdee31e76b4585ad7e926173f59dbf42a185adb"} Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.601590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerStarted","Data":"00621d1bc065fdc0c28f4a1b06c18987f0f613beee85ffa31504d8cf618eb8ff"} Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.629169 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nxffg" podStartSLOduration=2.103311525 podStartE2EDuration="4.629149194s" podCreationTimestamp="2025-12-03 17:09:37 +0000 UTC" firstStartedPulling="2025-12-03 17:09:38.568940313 +0000 UTC m=+833.770317174" lastFinishedPulling="2025-12-03 17:09:41.094777952 +0000 UTC m=+836.296154843" observedRunningTime="2025-12-03 17:09:41.626157645 +0000 UTC m=+836.827534506" watchObservedRunningTime="2025-12-03 17:09:41.629149194 +0000 UTC m=+836.830526055" Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.850643 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.979594 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-util\") pod \"ad1568bc-9c16-4426-b740-c6ac65af84c0\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.979788 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz9j7\" (UniqueName: \"kubernetes.io/projected/ad1568bc-9c16-4426-b740-c6ac65af84c0-kube-api-access-cz9j7\") pod \"ad1568bc-9c16-4426-b740-c6ac65af84c0\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.979857 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-bundle\") pod \"ad1568bc-9c16-4426-b740-c6ac65af84c0\" (UID: \"ad1568bc-9c16-4426-b740-c6ac65af84c0\") " Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.980504 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-bundle" (OuterVolumeSpecName: "bundle") pod "ad1568bc-9c16-4426-b740-c6ac65af84c0" (UID: "ad1568bc-9c16-4426-b740-c6ac65af84c0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.987658 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad1568bc-9c16-4426-b740-c6ac65af84c0-kube-api-access-cz9j7" (OuterVolumeSpecName: "kube-api-access-cz9j7") pod "ad1568bc-9c16-4426-b740-c6ac65af84c0" (UID: "ad1568bc-9c16-4426-b740-c6ac65af84c0"). InnerVolumeSpecName "kube-api-access-cz9j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:09:41 crc kubenswrapper[4758]: I1203 17:09:41.994531 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-util" (OuterVolumeSpecName: "util") pod "ad1568bc-9c16-4426-b740-c6ac65af84c0" (UID: "ad1568bc-9c16-4426-b740-c6ac65af84c0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:09:42 crc kubenswrapper[4758]: I1203 17:09:42.081719 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz9j7\" (UniqueName: \"kubernetes.io/projected/ad1568bc-9c16-4426-b740-c6ac65af84c0-kube-api-access-cz9j7\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:42 crc kubenswrapper[4758]: I1203 17:09:42.082011 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:42 crc kubenswrapper[4758]: I1203 17:09:42.082072 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad1568bc-9c16-4426-b740-c6ac65af84c0-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:42 crc kubenswrapper[4758]: I1203 17:09:42.611201 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" event={"ID":"ad1568bc-9c16-4426-b740-c6ac65af84c0","Type":"ContainerDied","Data":"b81791c3945f9ff7fb27793de8ab3534fe7e79bcb33037eff32a1315f2b12c79"} Dec 03 17:09:42 crc kubenswrapper[4758]: I1203 17:09:42.611266 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b81791c3945f9ff7fb27793de8ab3534fe7e79bcb33037eff32a1315f2b12c79" Dec 03 17:09:42 crc kubenswrapper[4758]: I1203 17:09:42.611233 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.622695 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs"] Dec 03 17:09:45 crc kubenswrapper[4758]: E1203 17:09:45.623267 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="extract" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.623282 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="extract" Dec 03 17:09:45 crc kubenswrapper[4758]: E1203 17:09:45.623295 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="util" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.623301 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="util" Dec 03 17:09:45 crc kubenswrapper[4758]: E1203 17:09:45.623316 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="pull" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.623322 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="pull" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.623411 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad1568bc-9c16-4426-b740-c6ac65af84c0" containerName="extract" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.623882 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.626113 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.626158 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.626385 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-64m2g" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.633612 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs"] Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.750312 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44vs2\" (UniqueName: \"kubernetes.io/projected/a1a68f23-f8d5-40c2-bca8-cf0f7ae19044-kube-api-access-44vs2\") pod \"nmstate-operator-5b5b58f5c8-wdgqs\" (UID: \"a1a68f23-f8d5-40c2-bca8-cf0f7ae19044\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.852638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44vs2\" (UniqueName: \"kubernetes.io/projected/a1a68f23-f8d5-40c2-bca8-cf0f7ae19044-kube-api-access-44vs2\") pod \"nmstate-operator-5b5b58f5c8-wdgqs\" (UID: \"a1a68f23-f8d5-40c2-bca8-cf0f7ae19044\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.881146 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44vs2\" (UniqueName: \"kubernetes.io/projected/a1a68f23-f8d5-40c2-bca8-cf0f7ae19044-kube-api-access-44vs2\") pod \"nmstate-operator-5b5b58f5c8-wdgqs\" (UID: \"a1a68f23-f8d5-40c2-bca8-cf0f7ae19044\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" Dec 03 17:09:45 crc kubenswrapper[4758]: I1203 17:09:45.944365 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" Dec 03 17:09:46 crc kubenswrapper[4758]: I1203 17:09:46.967411 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs"] Dec 03 17:09:47 crc kubenswrapper[4758]: I1203 17:09:47.420958 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:47 crc kubenswrapper[4758]: I1203 17:09:47.421437 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:47 crc kubenswrapper[4758]: I1203 17:09:47.469669 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:47 crc kubenswrapper[4758]: I1203 17:09:47.656302 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" event={"ID":"a1a68f23-f8d5-40c2-bca8-cf0f7ae19044","Type":"ContainerStarted","Data":"ff865877d39e41b0fc6a2f94395e63faebaa1b603fe37bddb53e3d14685a5fba"} Dec 03 17:09:47 crc kubenswrapper[4758]: I1203 17:09:47.730289 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:49 crc kubenswrapper[4758]: I1203 17:09:49.672747 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" event={"ID":"a1a68f23-f8d5-40c2-bca8-cf0f7ae19044","Type":"ContainerStarted","Data":"d143bf6a0787b089bb22e363167e237bbc3a0b1fead878cb378f8cdf96847850"} Dec 03 17:09:49 crc kubenswrapper[4758]: I1203 17:09:49.700882 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wdgqs" podStartSLOduration=2.818881251 podStartE2EDuration="4.700857723s" podCreationTimestamp="2025-12-03 17:09:45 +0000 UTC" firstStartedPulling="2025-12-03 17:09:46.98102538 +0000 UTC m=+842.182402241" lastFinishedPulling="2025-12-03 17:09:48.863001852 +0000 UTC m=+844.064378713" observedRunningTime="2025-12-03 17:09:49.695069943 +0000 UTC m=+844.896446814" watchObservedRunningTime="2025-12-03 17:09:49.700857723 +0000 UTC m=+844.902234584" Dec 03 17:09:50 crc kubenswrapper[4758]: I1203 17:09:50.074805 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxffg"] Dec 03 17:09:50 crc kubenswrapper[4758]: I1203 17:09:50.075195 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nxffg" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="registry-server" containerID="cri-o://00621d1bc065fdc0c28f4a1b06c18987f0f613beee85ffa31504d8cf618eb8ff" gracePeriod=2 Dec 03 17:09:51 crc kubenswrapper[4758]: I1203 17:09:51.688040 4758 generic.go:334] "Generic (PLEG): container finished" podID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerID="00621d1bc065fdc0c28f4a1b06c18987f0f613beee85ffa31504d8cf618eb8ff" exitCode=0 Dec 03 17:09:51 crc kubenswrapper[4758]: I1203 17:09:51.688127 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerDied","Data":"00621d1bc065fdc0c28f4a1b06c18987f0f613beee85ffa31504d8cf618eb8ff"} Dec 03 17:09:53 crc kubenswrapper[4758]: I1203 17:09:53.871716 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:53 crc kubenswrapper[4758]: I1203 17:09:53.992197 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-catalog-content\") pod \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " Dec 03 17:09:53 crc kubenswrapper[4758]: I1203 17:09:53.992316 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-utilities\") pod \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " Dec 03 17:09:53 crc kubenswrapper[4758]: I1203 17:09:53.992416 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcrlc\" (UniqueName: \"kubernetes.io/projected/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-kube-api-access-fcrlc\") pod \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\" (UID: \"fa036f9c-f46c-4045-9f0b-7c880b62f1f8\") " Dec 03 17:09:53 crc kubenswrapper[4758]: I1203 17:09:53.993505 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-utilities" (OuterVolumeSpecName: "utilities") pod "fa036f9c-f46c-4045-9f0b-7c880b62f1f8" (UID: "fa036f9c-f46c-4045-9f0b-7c880b62f1f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.000044 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-kube-api-access-fcrlc" (OuterVolumeSpecName: "kube-api-access-fcrlc") pod "fa036f9c-f46c-4045-9f0b-7c880b62f1f8" (UID: "fa036f9c-f46c-4045-9f0b-7c880b62f1f8"). InnerVolumeSpecName "kube-api-access-fcrlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.094369 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.094412 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcrlc\" (UniqueName: \"kubernetes.io/projected/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-kube-api-access-fcrlc\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.096316 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa036f9c-f46c-4045-9f0b-7c880b62f1f8" (UID: "fa036f9c-f46c-4045-9f0b-7c880b62f1f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.195562 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa036f9c-f46c-4045-9f0b-7c880b62f1f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.708188 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxffg" event={"ID":"fa036f9c-f46c-4045-9f0b-7c880b62f1f8","Type":"ContainerDied","Data":"e9bb0cdcd72cd82c1c80c35481a9d5285a102d6029ca50520e3bcd829548e188"} Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.708241 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxffg" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.708298 4758 scope.go:117] "RemoveContainer" containerID="00621d1bc065fdc0c28f4a1b06c18987f0f613beee85ffa31504d8cf618eb8ff" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.735020 4758 scope.go:117] "RemoveContainer" containerID="b04b5c1bdc46c6b766455c52a1c1439cb612a1b2bde98a3a2a3d590c5d8da9d3" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.753367 4758 scope.go:117] "RemoveContainer" containerID="fbe91367b18a2565736167f76c56489c069ad3c51e1c36746d92bc7b0ff322ca" Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.760747 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxffg"] Dec 03 17:09:54 crc kubenswrapper[4758]: I1203 17:09:54.767980 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nxffg"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.120914 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" path="/var/lib/kubelet/pods/fa036f9c-f46c-4045-9f0b-7c880b62f1f8/volumes" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.348788 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg"] Dec 03 17:09:55 crc kubenswrapper[4758]: E1203 17:09:55.349100 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="extract-utilities" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.349127 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="extract-utilities" Dec 03 17:09:55 crc kubenswrapper[4758]: E1203 17:09:55.349158 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="registry-server" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.349168 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="registry-server" Dec 03 17:09:55 crc kubenswrapper[4758]: E1203 17:09:55.349183 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="extract-content" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.349193 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="extract-content" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.349321 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa036f9c-f46c-4045-9f0b-7c880b62f1f8" containerName="registry-server" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.350151 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.355407 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.356756 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.360846 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.361056 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nn82k" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.370545 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.376405 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.381348 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-xfg7s"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.382463 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-dbus-socket\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516514 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-nmstate-lock\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516547 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7267aad3-2922-44b5-85b2-4c40d43c5849-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-tjzhr\" (UID: \"7267aad3-2922-44b5-85b2-4c40d43c5849\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516569 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-ovs-socket\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516604 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82nt9\" (UniqueName: \"kubernetes.io/projected/edb53861-39ab-4c5c-ac89-7cd726ce6dcc-kube-api-access-82nt9\") pod \"nmstate-metrics-7f946cbc9-c92pg\" (UID: \"edb53861-39ab-4c5c-ac89-7cd726ce6dcc\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbfnm\" (UniqueName: \"kubernetes.io/projected/dd83999a-a4d2-4880-ab08-2e73e2eddd21-kube-api-access-wbfnm\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.516705 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqcb6\" (UniqueName: \"kubernetes.io/projected/7267aad3-2922-44b5-85b2-4c40d43c5849-kube-api-access-qqcb6\") pod \"nmstate-webhook-5f6d4c5ccb-tjzhr\" (UID: \"7267aad3-2922-44b5-85b2-4c40d43c5849\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.522760 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.524530 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.527615 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.528026 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.528466 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vgfkh" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.532418 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619777 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eff08a5-e932-4488-9f0f-4a4584ec8c11-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619835 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4eff08a5-e932-4488-9f0f-4a4584ec8c11-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619880 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbfnm\" (UniqueName: \"kubernetes.io/projected/dd83999a-a4d2-4880-ab08-2e73e2eddd21-kube-api-access-wbfnm\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619905 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqcb6\" (UniqueName: \"kubernetes.io/projected/7267aad3-2922-44b5-85b2-4c40d43c5849-kube-api-access-qqcb6\") pod \"nmstate-webhook-5f6d4c5ccb-tjzhr\" (UID: \"7267aad3-2922-44b5-85b2-4c40d43c5849\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619933 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-dbus-socket\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619952 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-nmstate-lock\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7267aad3-2922-44b5-85b2-4c40d43c5849-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-tjzhr\" (UID: \"7267aad3-2922-44b5-85b2-4c40d43c5849\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.619985 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-ovs-socket\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.620007 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv72k\" (UniqueName: \"kubernetes.io/projected/4eff08a5-e932-4488-9f0f-4a4584ec8c11-kube-api-access-pv72k\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.620032 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82nt9\" (UniqueName: \"kubernetes.io/projected/edb53861-39ab-4c5c-ac89-7cd726ce6dcc-kube-api-access-82nt9\") pod \"nmstate-metrics-7f946cbc9-c92pg\" (UID: \"edb53861-39ab-4c5c-ac89-7cd726ce6dcc\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.620873 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-dbus-socket\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.620912 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-nmstate-lock\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.622042 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/dd83999a-a4d2-4880-ab08-2e73e2eddd21-ovs-socket\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.628419 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7267aad3-2922-44b5-85b2-4c40d43c5849-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-tjzhr\" (UID: \"7267aad3-2922-44b5-85b2-4c40d43c5849\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.638237 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqcb6\" (UniqueName: \"kubernetes.io/projected/7267aad3-2922-44b5-85b2-4c40d43c5849-kube-api-access-qqcb6\") pod \"nmstate-webhook-5f6d4c5ccb-tjzhr\" (UID: \"7267aad3-2922-44b5-85b2-4c40d43c5849\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.638905 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82nt9\" (UniqueName: \"kubernetes.io/projected/edb53861-39ab-4c5c-ac89-7cd726ce6dcc-kube-api-access-82nt9\") pod \"nmstate-metrics-7f946cbc9-c92pg\" (UID: \"edb53861-39ab-4c5c-ac89-7cd726ce6dcc\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.643410 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbfnm\" (UniqueName: \"kubernetes.io/projected/dd83999a-a4d2-4880-ab08-2e73e2eddd21-kube-api-access-wbfnm\") pod \"nmstate-handler-xfg7s\" (UID: \"dd83999a-a4d2-4880-ab08-2e73e2eddd21\") " pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.671953 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.678536 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5b96db6c87-4cd9b"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.679380 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.694183 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b96db6c87-4cd9b"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.694547 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.704514 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.721145 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eff08a5-e932-4488-9f0f-4a4584ec8c11-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.721198 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4eff08a5-e932-4488-9f0f-4a4584ec8c11-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.721266 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv72k\" (UniqueName: \"kubernetes.io/projected/4eff08a5-e932-4488-9f0f-4a4584ec8c11-kube-api-access-pv72k\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.723489 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4eff08a5-e932-4488-9f0f-4a4584ec8c11-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.729337 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eff08a5-e932-4488-9f0f-4a4584ec8c11-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.736968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv72k\" (UniqueName: \"kubernetes.io/projected/4eff08a5-e932-4488-9f0f-4a4584ec8c11-kube-api-access-pv72k\") pod \"nmstate-console-plugin-7fbb5f6569-cznmb\" (UID: \"4eff08a5-e932-4488-9f0f-4a4584ec8c11\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.826455 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-oauth-serving-cert\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.826969 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-service-ca\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.826995 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5a337395-8254-407b-8183-0200fe08a4e5-console-oauth-config\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.827010 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-trusted-ca-bundle\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.827042 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5a337395-8254-407b-8183-0200fe08a4e5-console-serving-cert\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.827063 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkg46\" (UniqueName: \"kubernetes.io/projected/5a337395-8254-407b-8183-0200fe08a4e5-kube-api-access-pkg46\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.827083 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-console-config\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.849531 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.916598 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg"] Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966170 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkg46\" (UniqueName: \"kubernetes.io/projected/5a337395-8254-407b-8183-0200fe08a4e5-kube-api-access-pkg46\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966218 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-console-config\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966257 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-oauth-serving-cert\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966304 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-service-ca\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5a337395-8254-407b-8183-0200fe08a4e5-console-oauth-config\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966347 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-trusted-ca-bundle\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.966363 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5a337395-8254-407b-8183-0200fe08a4e5-console-serving-cert\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.967867 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-oauth-serving-cert\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.968222 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-console-config\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: W1203 17:09:55.970452 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedb53861_39ab_4c5c_ac89_7cd726ce6dcc.slice/crio-eac85ea0716397f6bbc3263072ffdfac314578e21a7d04a690e1e319e8e68366 WatchSource:0}: Error finding container eac85ea0716397f6bbc3263072ffdfac314578e21a7d04a690e1e319e8e68366: Status 404 returned error can't find the container with id eac85ea0716397f6bbc3263072ffdfac314578e21a7d04a690e1e319e8e68366 Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.970586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-service-ca\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.970972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a337395-8254-407b-8183-0200fe08a4e5-trusted-ca-bundle\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.981799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5a337395-8254-407b-8183-0200fe08a4e5-console-oauth-config\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.983627 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5a337395-8254-407b-8183-0200fe08a4e5-console-serving-cert\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:55 crc kubenswrapper[4758]: I1203 17:09:55.987374 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkg46\" (UniqueName: \"kubernetes.io/projected/5a337395-8254-407b-8183-0200fe08a4e5-kube-api-access-pkg46\") pod \"console-5b96db6c87-4cd9b\" (UID: \"5a337395-8254-407b-8183-0200fe08a4e5\") " pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.027914 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr"] Dec 03 17:09:56 crc kubenswrapper[4758]: W1203 17:09:56.038004 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7267aad3_2922_44b5_85b2_4c40d43c5849.slice/crio-558479241dab372e26bfa767c39a6ec6e42c50f8256141b6a0a5068a7612c47c WatchSource:0}: Error finding container 558479241dab372e26bfa767c39a6ec6e42c50f8256141b6a0a5068a7612c47c: Status 404 returned error can't find the container with id 558479241dab372e26bfa767c39a6ec6e42c50f8256141b6a0a5068a7612c47c Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.057879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.115805 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb"] Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.305325 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b96db6c87-4cd9b"] Dec 03 17:09:56 crc kubenswrapper[4758]: W1203 17:09:56.312673 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a337395_8254_407b_8183_0200fe08a4e5.slice/crio-4ba5c57c891de6d8dd4e488b8d921fa9df8bd233665897a56f6c305ab4177461 WatchSource:0}: Error finding container 4ba5c57c891de6d8dd4e488b8d921fa9df8bd233665897a56f6c305ab4177461: Status 404 returned error can't find the container with id 4ba5c57c891de6d8dd4e488b8d921fa9df8bd233665897a56f6c305ab4177461 Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.744157 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" event={"ID":"7267aad3-2922-44b5-85b2-4c40d43c5849","Type":"ContainerStarted","Data":"558479241dab372e26bfa767c39a6ec6e42c50f8256141b6a0a5068a7612c47c"} Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.745832 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b96db6c87-4cd9b" event={"ID":"5a337395-8254-407b-8183-0200fe08a4e5","Type":"ContainerStarted","Data":"ae77a91282f8d284c12722ea31b6d1cb8e5297fa43a0ed8552b3c8bfccb16335"} Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.745882 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b96db6c87-4cd9b" event={"ID":"5a337395-8254-407b-8183-0200fe08a4e5","Type":"ContainerStarted","Data":"4ba5c57c891de6d8dd4e488b8d921fa9df8bd233665897a56f6c305ab4177461"} Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.747440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" event={"ID":"edb53861-39ab-4c5c-ac89-7cd726ce6dcc","Type":"ContainerStarted","Data":"eac85ea0716397f6bbc3263072ffdfac314578e21a7d04a690e1e319e8e68366"} Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.748704 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" event={"ID":"4eff08a5-e932-4488-9f0f-4a4584ec8c11","Type":"ContainerStarted","Data":"bc76158883f43b282dfb24a3bf6421c0cf76edf33c224dda2f3ca607117158d8"} Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.750554 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xfg7s" event={"ID":"dd83999a-a4d2-4880-ab08-2e73e2eddd21","Type":"ContainerStarted","Data":"a14fe201b4184c2de6fcd4edc9bc855d91b89a665f77716dace67dcdd5dface8"} Dec 03 17:09:56 crc kubenswrapper[4758]: I1203 17:09:56.768349 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5b96db6c87-4cd9b" podStartSLOduration=1.768327698 podStartE2EDuration="1.768327698s" podCreationTimestamp="2025-12-03 17:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:09:56.767274528 +0000 UTC m=+851.968651399" watchObservedRunningTime="2025-12-03 17:09:56.768327698 +0000 UTC m=+851.969704559" Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.773657 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xfg7s" event={"ID":"dd83999a-a4d2-4880-ab08-2e73e2eddd21","Type":"ContainerStarted","Data":"94cda9eb12ce8b1bfe2ae5858269e6fc654317ac47bd654b963af98d1f4b8982"} Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.777101 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.777222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" event={"ID":"7267aad3-2922-44b5-85b2-4c40d43c5849","Type":"ContainerStarted","Data":"70a6e31256b3a80bcae2ed632c518181160699cc6fa099024c8c0ec6e429623f"} Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.777479 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.779860 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" event={"ID":"edb53861-39ab-4c5c-ac89-7cd726ce6dcc","Type":"ContainerStarted","Data":"4b70e115ba9df7ff650fd124d4552b9dda3003a9ae5e1589a0a78b54c0fc985d"} Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.781539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" event={"ID":"4eff08a5-e932-4488-9f0f-4a4584ec8c11","Type":"ContainerStarted","Data":"d23767b1995ce8dd9dfd1887566ea2bdc2c470825e8e7559751a7557dc4697f3"} Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.796541 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-xfg7s" podStartSLOduration=1.394753469 podStartE2EDuration="4.796496588s" podCreationTimestamp="2025-12-03 17:09:55 +0000 UTC" firstStartedPulling="2025-12-03 17:09:55.745275857 +0000 UTC m=+850.946652718" lastFinishedPulling="2025-12-03 17:09:59.147018976 +0000 UTC m=+854.348395837" observedRunningTime="2025-12-03 17:09:59.792409295 +0000 UTC m=+854.993786166" watchObservedRunningTime="2025-12-03 17:09:59.796496588 +0000 UTC m=+854.997873459" Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.819913 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" podStartSLOduration=1.714177523 podStartE2EDuration="4.819883953s" podCreationTimestamp="2025-12-03 17:09:55 +0000 UTC" firstStartedPulling="2025-12-03 17:09:56.039761023 +0000 UTC m=+851.241137884" lastFinishedPulling="2025-12-03 17:09:59.145467453 +0000 UTC m=+854.346844314" observedRunningTime="2025-12-03 17:09:59.815881813 +0000 UTC m=+855.017258694" watchObservedRunningTime="2025-12-03 17:09:59.819883953 +0000 UTC m=+855.021260814" Dec 03 17:09:59 crc kubenswrapper[4758]: I1203 17:09:59.833451 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cznmb" podStartSLOduration=1.8328582679999998 podStartE2EDuration="4.833422977s" podCreationTimestamp="2025-12-03 17:09:55 +0000 UTC" firstStartedPulling="2025-12-03 17:09:56.127967747 +0000 UTC m=+851.329344608" lastFinishedPulling="2025-12-03 17:09:59.128532456 +0000 UTC m=+854.329909317" observedRunningTime="2025-12-03 17:09:59.832512842 +0000 UTC m=+855.033889733" watchObservedRunningTime="2025-12-03 17:09:59.833422977 +0000 UTC m=+855.034799838" Dec 03 17:10:02 crc kubenswrapper[4758]: I1203 17:10:02.828768 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" event={"ID":"edb53861-39ab-4c5c-ac89-7cd726ce6dcc","Type":"ContainerStarted","Data":"4235571f9aecfa36623e6c0e2c5790ce8698ce6714ae6cbf79e2807f3a89bf8d"} Dec 03 17:10:02 crc kubenswrapper[4758]: I1203 17:10:02.850330 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-c92pg" podStartSLOduration=1.900445703 podStartE2EDuration="7.850285765s" podCreationTimestamp="2025-12-03 17:09:55 +0000 UTC" firstStartedPulling="2025-12-03 17:09:55.976195879 +0000 UTC m=+851.177572740" lastFinishedPulling="2025-12-03 17:10:01.926035941 +0000 UTC m=+857.127412802" observedRunningTime="2025-12-03 17:10:02.842715097 +0000 UTC m=+858.044091958" watchObservedRunningTime="2025-12-03 17:10:02.850285765 +0000 UTC m=+858.051662626" Dec 03 17:10:05 crc kubenswrapper[4758]: I1203 17:10:05.733043 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-xfg7s" Dec 03 17:10:06 crc kubenswrapper[4758]: I1203 17:10:06.059497 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:10:06 crc kubenswrapper[4758]: I1203 17:10:06.059586 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:10:06 crc kubenswrapper[4758]: I1203 17:10:06.065610 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:10:06 crc kubenswrapper[4758]: I1203 17:10:06.858701 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5b96db6c87-4cd9b" Dec 03 17:10:06 crc kubenswrapper[4758]: I1203 17:10:06.912784 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kzv6z"] Dec 03 17:10:15 crc kubenswrapper[4758]: I1203 17:10:15.703774 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tjzhr" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.033998 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h"] Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.037229 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.039712 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.055375 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h"] Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.134359 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.134495 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqm8g\" (UniqueName: \"kubernetes.io/projected/71ef144e-1cc4-4f43-bf14-16237cf240b4-kube-api-access-rqm8g\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.134583 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.236458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.236564 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqm8g\" (UniqueName: \"kubernetes.io/projected/71ef144e-1cc4-4f43-bf14-16237cf240b4-kube-api-access-rqm8g\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.236616 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.237178 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.237355 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.262047 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqm8g\" (UniqueName: \"kubernetes.io/projected/71ef144e-1cc4-4f43-bf14-16237cf240b4-kube-api-access-rqm8g\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.347704 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-kzv6z" podUID="a32f32fc-e1aa-402c-93ec-f73446955bd0" containerName="console" containerID="cri-o://7a2835a436f3300583867ea64cb21b13fed5b722bdd540d7a22e4e65d4355463" gracePeriod=15 Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.357474 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:32 crc kubenswrapper[4758]: I1203 17:10:32.586421 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h"] Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.039165 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" event={"ID":"71ef144e-1cc4-4f43-bf14-16237cf240b4","Type":"ContainerStarted","Data":"913a67f4f36efbdbafb786ecd44421b4459bc4b6be755b061741606dd7f95b98"} Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.039214 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" event={"ID":"71ef144e-1cc4-4f43-bf14-16237cf240b4","Type":"ContainerStarted","Data":"8e22d698ad2b9ca55adfcf5dc4f6dfc2a4707a427d319ee7d560975aa332ea73"} Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.045360 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kzv6z_a32f32fc-e1aa-402c-93ec-f73446955bd0/console/0.log" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.045421 4758 generic.go:334] "Generic (PLEG): container finished" podID="a32f32fc-e1aa-402c-93ec-f73446955bd0" containerID="7a2835a436f3300583867ea64cb21b13fed5b722bdd540d7a22e4e65d4355463" exitCode=2 Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.045465 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kzv6z" event={"ID":"a32f32fc-e1aa-402c-93ec-f73446955bd0","Type":"ContainerDied","Data":"7a2835a436f3300583867ea64cb21b13fed5b722bdd540d7a22e4e65d4355463"} Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.373523 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kzv6z_a32f32fc-e1aa-402c-93ec-f73446955bd0/console/0.log" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.373994 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.456428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-oauth-serving-cert\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.456539 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-config\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.456601 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-trusted-ca-bundle\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.456618 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-oauth-config\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.456657 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-service-ca\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.457769 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-service-ca" (OuterVolumeSpecName: "service-ca") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.457874 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vt4s\" (UniqueName: \"kubernetes.io/projected/a32f32fc-e1aa-402c-93ec-f73446955bd0-kube-api-access-2vt4s\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.457791 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.457904 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-serving-cert\") pod \"a32f32fc-e1aa-402c-93ec-f73446955bd0\" (UID: \"a32f32fc-e1aa-402c-93ec-f73446955bd0\") " Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.457832 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.458353 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.458370 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.458380 4758 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.458401 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-config" (OuterVolumeSpecName: "console-config") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.464592 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.464787 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32f32fc-e1aa-402c-93ec-f73446955bd0-kube-api-access-2vt4s" (OuterVolumeSpecName: "kube-api-access-2vt4s") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "kube-api-access-2vt4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.469135 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a32f32fc-e1aa-402c-93ec-f73446955bd0" (UID: "a32f32fc-e1aa-402c-93ec-f73446955bd0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.560480 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vt4s\" (UniqueName: \"kubernetes.io/projected/a32f32fc-e1aa-402c-93ec-f73446955bd0-kube-api-access-2vt4s\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.560541 4758 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.560552 4758 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:33 crc kubenswrapper[4758]: I1203 17:10:33.560562 4758 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a32f32fc-e1aa-402c-93ec-f73446955bd0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.053407 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kzv6z_a32f32fc-e1aa-402c-93ec-f73446955bd0/console/0.log" Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.053512 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kzv6z" event={"ID":"a32f32fc-e1aa-402c-93ec-f73446955bd0","Type":"ContainerDied","Data":"3d18e8d674c5a8ab06381aa52458283e147d63ef2bb4b4301bee2b2160ccee05"} Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.053638 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kzv6z" Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.053650 4758 scope.go:117] "RemoveContainer" containerID="7a2835a436f3300583867ea64cb21b13fed5b722bdd540d7a22e4e65d4355463" Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.055843 4758 generic.go:334] "Generic (PLEG): container finished" podID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerID="913a67f4f36efbdbafb786ecd44421b4459bc4b6be755b061741606dd7f95b98" exitCode=0 Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.055922 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" event={"ID":"71ef144e-1cc4-4f43-bf14-16237cf240b4","Type":"ContainerDied","Data":"913a67f4f36efbdbafb786ecd44421b4459bc4b6be755b061741606dd7f95b98"} Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.099579 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kzv6z"] Dec 03 17:10:34 crc kubenswrapper[4758]: I1203 17:10:34.102972 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-kzv6z"] Dec 03 17:10:35 crc kubenswrapper[4758]: I1203 17:10:35.127459 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32f32fc-e1aa-402c-93ec-f73446955bd0" path="/var/lib/kubelet/pods/a32f32fc-e1aa-402c-93ec-f73446955bd0/volumes" Dec 03 17:10:36 crc kubenswrapper[4758]: I1203 17:10:36.075457 4758 generic.go:334] "Generic (PLEG): container finished" podID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerID="0e670cd646b1b51e73e652dbe26ae8ac8a89278b2856b6e1c4034bbf7b4092cd" exitCode=0 Dec 03 17:10:36 crc kubenswrapper[4758]: I1203 17:10:36.075520 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" event={"ID":"71ef144e-1cc4-4f43-bf14-16237cf240b4","Type":"ContainerDied","Data":"0e670cd646b1b51e73e652dbe26ae8ac8a89278b2856b6e1c4034bbf7b4092cd"} Dec 03 17:10:37 crc kubenswrapper[4758]: I1203 17:10:37.085503 4758 generic.go:334] "Generic (PLEG): container finished" podID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerID="3273cd01f1dcc5d522932d2f1ef4ec322948c5ff5e11264946819e6b0346c79d" exitCode=0 Dec 03 17:10:37 crc kubenswrapper[4758]: I1203 17:10:37.085576 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" event={"ID":"71ef144e-1cc4-4f43-bf14-16237cf240b4","Type":"ContainerDied","Data":"3273cd01f1dcc5d522932d2f1ef4ec322948c5ff5e11264946819e6b0346c79d"} Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.342210 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.440204 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-util\") pod \"71ef144e-1cc4-4f43-bf14-16237cf240b4\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.440417 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqm8g\" (UniqueName: \"kubernetes.io/projected/71ef144e-1cc4-4f43-bf14-16237cf240b4-kube-api-access-rqm8g\") pod \"71ef144e-1cc4-4f43-bf14-16237cf240b4\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.440456 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-bundle\") pod \"71ef144e-1cc4-4f43-bf14-16237cf240b4\" (UID: \"71ef144e-1cc4-4f43-bf14-16237cf240b4\") " Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.442609 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-bundle" (OuterVolumeSpecName: "bundle") pod "71ef144e-1cc4-4f43-bf14-16237cf240b4" (UID: "71ef144e-1cc4-4f43-bf14-16237cf240b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.448977 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71ef144e-1cc4-4f43-bf14-16237cf240b4-kube-api-access-rqm8g" (OuterVolumeSpecName: "kube-api-access-rqm8g") pod "71ef144e-1cc4-4f43-bf14-16237cf240b4" (UID: "71ef144e-1cc4-4f43-bf14-16237cf240b4"). InnerVolumeSpecName "kube-api-access-rqm8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.542188 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.542225 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqm8g\" (UniqueName: \"kubernetes.io/projected/71ef144e-1cc4-4f43-bf14-16237cf240b4-kube-api-access-rqm8g\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.582352 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-util" (OuterVolumeSpecName: "util") pod "71ef144e-1cc4-4f43-bf14-16237cf240b4" (UID: "71ef144e-1cc4-4f43-bf14-16237cf240b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:10:38 crc kubenswrapper[4758]: I1203 17:10:38.643719 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71ef144e-1cc4-4f43-bf14-16237cf240b4-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:10:39 crc kubenswrapper[4758]: I1203 17:10:39.104225 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" event={"ID":"71ef144e-1cc4-4f43-bf14-16237cf240b4","Type":"ContainerDied","Data":"8e22d698ad2b9ca55adfcf5dc4f6dfc2a4707a427d319ee7d560975aa332ea73"} Dec 03 17:10:39 crc kubenswrapper[4758]: I1203 17:10:39.104337 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e22d698ad2b9ca55adfcf5dc4f6dfc2a4707a427d319ee7d560975aa332ea73" Dec 03 17:10:39 crc kubenswrapper[4758]: I1203 17:10:39.104431 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.045200 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn"] Dec 03 17:10:47 crc kubenswrapper[4758]: E1203 17:10:47.046170 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="util" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046185 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="util" Dec 03 17:10:47 crc kubenswrapper[4758]: E1203 17:10:47.046201 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="pull" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046209 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="pull" Dec 03 17:10:47 crc kubenswrapper[4758]: E1203 17:10:47.046221 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32f32fc-e1aa-402c-93ec-f73446955bd0" containerName="console" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046227 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32f32fc-e1aa-402c-93ec-f73446955bd0" containerName="console" Dec 03 17:10:47 crc kubenswrapper[4758]: E1203 17:10:47.046236 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="extract" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046243 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="extract" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046347 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71ef144e-1cc4-4f43-bf14-16237cf240b4" containerName="extract" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046363 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32f32fc-e1aa-402c-93ec-f73446955bd0" containerName="console" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.046904 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.052741 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.052745 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.052892 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.053450 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.053615 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zr5xm" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.063196 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-apiservice-cert\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.063282 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hxcb\" (UniqueName: \"kubernetes.io/projected/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-kube-api-access-7hxcb\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.063323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-webhook-cert\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.080467 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn"] Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.164043 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hxcb\" (UniqueName: \"kubernetes.io/projected/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-kube-api-access-7hxcb\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.164095 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-webhook-cert\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.164171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-apiservice-cert\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.186030 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-apiservice-cert\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.186082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-webhook-cert\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.192318 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hxcb\" (UniqueName: \"kubernetes.io/projected/3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e-kube-api-access-7hxcb\") pod \"metallb-operator-controller-manager-77cf4cbccf-5wfbn\" (UID: \"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e\") " pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.316233 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r"] Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.316992 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.321917 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.322294 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.322999 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-69v4m" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.350311 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r"] Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.364189 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.466761 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cfc8f20-5774-4fb3-ab06-179070a49c4c-webhook-cert\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.467312 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cfc8f20-5774-4fb3-ab06-179070a49c4c-apiservice-cert\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.467337 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2pzz\" (UniqueName: \"kubernetes.io/projected/9cfc8f20-5774-4fb3-ab06-179070a49c4c-kube-api-access-s2pzz\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.569209 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cfc8f20-5774-4fb3-ab06-179070a49c4c-webhook-cert\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.569313 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cfc8f20-5774-4fb3-ab06-179070a49c4c-apiservice-cert\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.569343 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2pzz\" (UniqueName: \"kubernetes.io/projected/9cfc8f20-5774-4fb3-ab06-179070a49c4c-kube-api-access-s2pzz\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.586214 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cfc8f20-5774-4fb3-ab06-179070a49c4c-webhook-cert\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.594531 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cfc8f20-5774-4fb3-ab06-179070a49c4c-apiservice-cert\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.613516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2pzz\" (UniqueName: \"kubernetes.io/projected/9cfc8f20-5774-4fb3-ab06-179070a49c4c-kube-api-access-s2pzz\") pod \"metallb-operator-webhook-server-547d88fbfd-ns59r\" (UID: \"9cfc8f20-5774-4fb3-ab06-179070a49c4c\") " pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.633082 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.760198 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn"] Dec 03 17:10:47 crc kubenswrapper[4758]: W1203 17:10:47.774948 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b2b2fc1_2f85_4c3b_956c_27378e0c4e5e.slice/crio-be2e42a2e4c10b0e6483bfea7b3436dffc80196005a18cd9ec7a14859a8b5842 WatchSource:0}: Error finding container be2e42a2e4c10b0e6483bfea7b3436dffc80196005a18cd9ec7a14859a8b5842: Status 404 returned error can't find the container with id be2e42a2e4c10b0e6483bfea7b3436dffc80196005a18cd9ec7a14859a8b5842 Dec 03 17:10:47 crc kubenswrapper[4758]: I1203 17:10:47.970866 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r"] Dec 03 17:10:47 crc kubenswrapper[4758]: W1203 17:10:47.979107 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cfc8f20_5774_4fb3_ab06_179070a49c4c.slice/crio-40287ed543612cd758c07612578697d4d623f83094bbb8e66326559f3ffea7b1 WatchSource:0}: Error finding container 40287ed543612cd758c07612578697d4d623f83094bbb8e66326559f3ffea7b1: Status 404 returned error can't find the container with id 40287ed543612cd758c07612578697d4d623f83094bbb8e66326559f3ffea7b1 Dec 03 17:10:48 crc kubenswrapper[4758]: I1203 17:10:48.154173 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" event={"ID":"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e","Type":"ContainerStarted","Data":"be2e42a2e4c10b0e6483bfea7b3436dffc80196005a18cd9ec7a14859a8b5842"} Dec 03 17:10:48 crc kubenswrapper[4758]: I1203 17:10:48.155587 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" event={"ID":"9cfc8f20-5774-4fb3-ab06-179070a49c4c","Type":"ContainerStarted","Data":"40287ed543612cd758c07612578697d4d623f83094bbb8e66326559f3ffea7b1"} Dec 03 17:10:54 crc kubenswrapper[4758]: I1203 17:10:54.198791 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" event={"ID":"9cfc8f20-5774-4fb3-ab06-179070a49c4c","Type":"ContainerStarted","Data":"f10de8ca4fa77955b459379f3b949f08d1c424dfaaf9623aed8a0ea3d14698c0"} Dec 03 17:10:54 crc kubenswrapper[4758]: I1203 17:10:54.200644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" event={"ID":"3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e","Type":"ContainerStarted","Data":"f1a52423c84e85d1d3d8f44ac3198b148fff4c24d7d63185b2af93e5557eaef1"} Dec 03 17:10:54 crc kubenswrapper[4758]: I1203 17:10:54.200769 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:10:54 crc kubenswrapper[4758]: I1203 17:10:54.200881 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:10:54 crc kubenswrapper[4758]: I1203 17:10:54.222895 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" podStartSLOduration=2.052004496 podStartE2EDuration="7.22287211s" podCreationTimestamp="2025-12-03 17:10:47 +0000 UTC" firstStartedPulling="2025-12-03 17:10:47.982903465 +0000 UTC m=+903.184280326" lastFinishedPulling="2025-12-03 17:10:53.153771089 +0000 UTC m=+908.355147940" observedRunningTime="2025-12-03 17:10:54.220867073 +0000 UTC m=+909.422243934" watchObservedRunningTime="2025-12-03 17:10:54.22287211 +0000 UTC m=+909.424248971" Dec 03 17:10:54 crc kubenswrapper[4758]: I1203 17:10:54.272344 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" podStartSLOduration=1.908229581 podStartE2EDuration="7.272318547s" podCreationTimestamp="2025-12-03 17:10:47 +0000 UTC" firstStartedPulling="2025-12-03 17:10:47.780567801 +0000 UTC m=+902.981944662" lastFinishedPulling="2025-12-03 17:10:53.144656767 +0000 UTC m=+908.346033628" observedRunningTime="2025-12-03 17:10:54.26791938 +0000 UTC m=+909.469296251" watchObservedRunningTime="2025-12-03 17:10:54.272318547 +0000 UTC m=+909.473695408" Dec 03 17:11:07 crc kubenswrapper[4758]: I1203 17:11:07.709877 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-547d88fbfd-ns59r" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.034174 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sxzdg"] Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.035318 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.045479 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sxzdg"] Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.084801 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-catalog-content\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.084866 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d868d\" (UniqueName: \"kubernetes.io/projected/b3b8ac8b-1223-4e5b-8435-6f543ede9341-kube-api-access-d868d\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.084907 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-utilities\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.185990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-catalog-content\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.186072 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d868d\" (UniqueName: \"kubernetes.io/projected/b3b8ac8b-1223-4e5b-8435-6f543ede9341-kube-api-access-d868d\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.186114 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-utilities\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.186609 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-catalog-content\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.186698 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-utilities\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.209884 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d868d\" (UniqueName: \"kubernetes.io/projected/b3b8ac8b-1223-4e5b-8435-6f543ede9341-kube-api-access-d868d\") pod \"certified-operators-sxzdg\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.393812 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:08 crc kubenswrapper[4758]: I1203 17:11:08.726294 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sxzdg"] Dec 03 17:11:09 crc kubenswrapper[4758]: I1203 17:11:09.311011 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxzdg" event={"ID":"b3b8ac8b-1223-4e5b-8435-6f543ede9341","Type":"ContainerStarted","Data":"d3948678b6587e0762114a4fb3730b10e1dde8b157de04c17dfb2e8b26dede1b"} Dec 03 17:11:11 crc kubenswrapper[4758]: I1203 17:11:11.394551 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:11:11 crc kubenswrapper[4758]: I1203 17:11:11.395080 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:11:12 crc kubenswrapper[4758]: I1203 17:11:12.333211 4758 generic.go:334] "Generic (PLEG): container finished" podID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerID="78dab2dd331fae43b01c253dabbff4a93d822bc60126dc00b04eaae42aab76d9" exitCode=0 Dec 03 17:11:12 crc kubenswrapper[4758]: I1203 17:11:12.333320 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxzdg" event={"ID":"b3b8ac8b-1223-4e5b-8435-6f543ede9341","Type":"ContainerDied","Data":"78dab2dd331fae43b01c253dabbff4a93d822bc60126dc00b04eaae42aab76d9"} Dec 03 17:11:13 crc kubenswrapper[4758]: I1203 17:11:13.343359 4758 generic.go:334] "Generic (PLEG): container finished" podID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerID="dafed731dfaa6e30e88ea170cfaf3ed2159ed88d88c0ddfe13f52584d07bde06" exitCode=0 Dec 03 17:11:13 crc kubenswrapper[4758]: I1203 17:11:13.343877 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxzdg" event={"ID":"b3b8ac8b-1223-4e5b-8435-6f543ede9341","Type":"ContainerDied","Data":"dafed731dfaa6e30e88ea170cfaf3ed2159ed88d88c0ddfe13f52584d07bde06"} Dec 03 17:11:15 crc kubenswrapper[4758]: I1203 17:11:15.358701 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxzdg" event={"ID":"b3b8ac8b-1223-4e5b-8435-6f543ede9341","Type":"ContainerStarted","Data":"c13229032f889eb33baff056451eee10715e89c0af754666c8df0dac41cb79b0"} Dec 03 17:11:15 crc kubenswrapper[4758]: I1203 17:11:15.381078 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sxzdg" podStartSLOduration=4.910221685 podStartE2EDuration="7.381054455s" podCreationTimestamp="2025-12-03 17:11:08 +0000 UTC" firstStartedPulling="2025-12-03 17:11:12.336122152 +0000 UTC m=+927.537499013" lastFinishedPulling="2025-12-03 17:11:14.806954932 +0000 UTC m=+930.008331783" observedRunningTime="2025-12-03 17:11:15.379935683 +0000 UTC m=+930.581312544" watchObservedRunningTime="2025-12-03 17:11:15.381054455 +0000 UTC m=+930.582431326" Dec 03 17:11:18 crc kubenswrapper[4758]: I1203 17:11:18.394904 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:18 crc kubenswrapper[4758]: I1203 17:11:18.395420 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:18 crc kubenswrapper[4758]: I1203 17:11:18.465663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.125970 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ws8lt"] Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.127931 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.135778 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws8lt"] Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.178341 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-catalog-content\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.178488 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-utilities\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.178577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9p49\" (UniqueName: \"kubernetes.io/projected/5b5a2278-6648-4466-8d43-c747ca02d55a-kube-api-access-d9p49\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.279438 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9p49\" (UniqueName: \"kubernetes.io/projected/5b5a2278-6648-4466-8d43-c747ca02d55a-kube-api-access-d9p49\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.279513 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-catalog-content\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.279574 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-utilities\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.280198 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-utilities\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.280385 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-catalog-content\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.299895 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9p49\" (UniqueName: \"kubernetes.io/projected/5b5a2278-6648-4466-8d43-c747ca02d55a-kube-api-access-d9p49\") pod \"redhat-marketplace-ws8lt\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.445391 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:24 crc kubenswrapper[4758]: I1203 17:11:24.748794 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws8lt"] Dec 03 17:11:25 crc kubenswrapper[4758]: I1203 17:11:25.424215 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerStarted","Data":"4ba878c2d1a080d75a922d04b08af26c1e26b36883cd406ff3579aa55d20dd93"} Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.367932 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-77cf4cbccf-5wfbn" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.512057 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cftwk"] Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.514207 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.529159 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cftwk"] Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.534187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-catalog-content\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.534283 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks4ws\" (UniqueName: \"kubernetes.io/projected/de9ffab2-2a7e-4146-96bf-1a97167257d6-kube-api-access-ks4ws\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.534393 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-utilities\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.635757 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-catalog-content\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.635810 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks4ws\" (UniqueName: \"kubernetes.io/projected/de9ffab2-2a7e-4146-96bf-1a97167257d6-kube-api-access-ks4ws\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.635865 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-utilities\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.636433 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-catalog-content\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.636500 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-utilities\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.656657 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks4ws\" (UniqueName: \"kubernetes.io/projected/de9ffab2-2a7e-4146-96bf-1a97167257d6-kube-api-access-ks4ws\") pod \"community-operators-cftwk\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:27 crc kubenswrapper[4758]: I1203 17:11:27.832260 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.262534 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-gxqxh"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.266698 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.273623 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.273774 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.274793 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.275760 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.283114 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.285445 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-f2sjb" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.286307 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cftwk"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.325473 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349539 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-reloader\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349594 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-sockets\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349622 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2psg\" (UniqueName: \"kubernetes.io/projected/272d2f0a-3217-4e67-82f3-43dc7aa334f5-kube-api-access-z2psg\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349653 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa19d1a0-7212-4b46-8877-9e4e09f1f775-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nqt6f\" (UID: \"fa19d1a0-7212-4b46-8877-9e4e09f1f775\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349673 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-startup\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349714 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics-certs\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349735 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-conf\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349761 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.349784 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6n5t\" (UniqueName: \"kubernetes.io/projected/fa19d1a0-7212-4b46-8877-9e4e09f1f775-kube-api-access-p6n5t\") pod \"frr-k8s-webhook-server-7fcb986d4-nqt6f\" (UID: \"fa19d1a0-7212-4b46-8877-9e4e09f1f775\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.448634 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerStarted","Data":"a6764f269385ee4b10a43b5243f636c732d6afea93998e29ed7930c3d827e31c"} Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451122 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451206 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6n5t\" (UniqueName: \"kubernetes.io/projected/fa19d1a0-7212-4b46-8877-9e4e09f1f775-kube-api-access-p6n5t\") pod \"frr-k8s-webhook-server-7fcb986d4-nqt6f\" (UID: \"fa19d1a0-7212-4b46-8877-9e4e09f1f775\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451280 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-reloader\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451365 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-sockets\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451399 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2psg\" (UniqueName: \"kubernetes.io/projected/272d2f0a-3217-4e67-82f3-43dc7aa334f5-kube-api-access-z2psg\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451459 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa19d1a0-7212-4b46-8877-9e4e09f1f775-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nqt6f\" (UID: \"fa19d1a0-7212-4b46-8877-9e4e09f1f775\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451482 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-startup\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451535 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics-certs\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.451563 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-conf\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.452215 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-sockets\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.452754 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.453345 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-reloader\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: E1203 17:11:28.453786 4758 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 17:11:28 crc kubenswrapper[4758]: E1203 17:11:28.453840 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics-certs podName:272d2f0a-3217-4e67-82f3-43dc7aa334f5 nodeName:}" failed. No retries permitted until 2025-12-03 17:11:28.953818898 +0000 UTC m=+944.155195759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics-certs") pod "frr-k8s-gxqxh" (UID: "272d2f0a-3217-4e67-82f3-43dc7aa334f5") : secret "frr-k8s-certs-secret" not found Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.454521 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-startup\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.455193 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/272d2f0a-3217-4e67-82f3-43dc7aa334f5-frr-conf\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.466426 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa19d1a0-7212-4b46-8877-9e4e09f1f775-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nqt6f\" (UID: \"fa19d1a0-7212-4b46-8877-9e4e09f1f775\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.487436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2psg\" (UniqueName: \"kubernetes.io/projected/272d2f0a-3217-4e67-82f3-43dc7aa334f5-kube-api-access-z2psg\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.488193 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6n5t\" (UniqueName: \"kubernetes.io/projected/fa19d1a0-7212-4b46-8877-9e4e09f1f775-kube-api-access-p6n5t\") pod \"frr-k8s-webhook-server-7fcb986d4-nqt6f\" (UID: \"fa19d1a0-7212-4b46-8877-9e4e09f1f775\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.497274 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zb4zw"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.499469 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.504230 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5rznf" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.504865 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.508089 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.509962 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.528489 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-vpzjk"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.533374 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.537362 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.537420 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553594 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553653 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01bb709f-7c93-45e7-b900-802665c0225f-metrics-certs\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553697 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-metrics-certs\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqbjz\" (UniqueName: \"kubernetes.io/projected/01bb709f-7c93-45e7-b900-802665c0225f-kube-api-access-cqbjz\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553752 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj7rf\" (UniqueName: \"kubernetes.io/projected/56cd1dfc-f881-441e-b231-d2783ddcf0f9-kube-api-access-qj7rf\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553787 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01bb709f-7c93-45e7-b900-802665c0225f-cert\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.553820 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/56cd1dfc-f881-441e-b231-d2783ddcf0f9-metallb-excludel2\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.564724 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-vpzjk"] Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655466 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01bb709f-7c93-45e7-b900-802665c0225f-metrics-certs\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655522 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-metrics-certs\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqbjz\" (UniqueName: \"kubernetes.io/projected/01bb709f-7c93-45e7-b900-802665c0225f-kube-api-access-cqbjz\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655580 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj7rf\" (UniqueName: \"kubernetes.io/projected/56cd1dfc-f881-441e-b231-d2783ddcf0f9-kube-api-access-qj7rf\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655623 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01bb709f-7c93-45e7-b900-802665c0225f-cert\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655662 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/56cd1dfc-f881-441e-b231-d2783ddcf0f9-metallb-excludel2\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.655769 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: E1203 17:11:28.655936 4758 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 17:11:28 crc kubenswrapper[4758]: E1203 17:11:28.655998 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist podName:56cd1dfc-f881-441e-b231-d2783ddcf0f9 nodeName:}" failed. No retries permitted until 2025-12-03 17:11:29.155978508 +0000 UTC m=+944.357355379 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist") pod "speaker-zb4zw" (UID: "56cd1dfc-f881-441e-b231-d2783ddcf0f9") : secret "metallb-memberlist" not found Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.658935 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.659628 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/56cd1dfc-f881-441e-b231-d2783ddcf0f9-metallb-excludel2\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.661326 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01bb709f-7c93-45e7-b900-802665c0225f-cert\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.667821 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01bb709f-7c93-45e7-b900-802665c0225f-metrics-certs\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.667895 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-metrics-certs\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.685537 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj7rf\" (UniqueName: \"kubernetes.io/projected/56cd1dfc-f881-441e-b231-d2783ddcf0f9-kube-api-access-qj7rf\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.701857 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqbjz\" (UniqueName: \"kubernetes.io/projected/01bb709f-7c93-45e7-b900-802665c0225f-kube-api-access-cqbjz\") pod \"controller-f8648f98b-vpzjk\" (UID: \"01bb709f-7c93-45e7-b900-802665c0225f\") " pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.861445 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.959942 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics-certs\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:28 crc kubenswrapper[4758]: I1203 17:11:28.967061 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/272d2f0a-3217-4e67-82f3-43dc7aa334f5-metrics-certs\") pod \"frr-k8s-gxqxh\" (UID: \"272d2f0a-3217-4e67-82f3-43dc7aa334f5\") " pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.007044 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f"] Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.163362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:29 crc kubenswrapper[4758]: E1203 17:11:29.163992 4758 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 17:11:29 crc kubenswrapper[4758]: E1203 17:11:29.164146 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist podName:56cd1dfc-f881-441e-b231-d2783ddcf0f9 nodeName:}" failed. No retries permitted until 2025-12-03 17:11:30.164084082 +0000 UTC m=+945.365461123 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist") pod "speaker-zb4zw" (UID: "56cd1dfc-f881-441e-b231-d2783ddcf0f9") : secret "metallb-memberlist" not found Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.189356 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-vpzjk"] Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.228799 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:29 crc kubenswrapper[4758]: W1203 17:11:29.241871 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01bb709f_7c93_45e7_b900_802665c0225f.slice/crio-682e5d147d28fd0a49a14821c981d2edda88646f4847a2eab128c899b4592ca5 WatchSource:0}: Error finding container 682e5d147d28fd0a49a14821c981d2edda88646f4847a2eab128c899b4592ca5: Status 404 returned error can't find the container with id 682e5d147d28fd0a49a14821c981d2edda88646f4847a2eab128c899b4592ca5 Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.456404 4758 generic.go:334] "Generic (PLEG): container finished" podID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerID="ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93" exitCode=0 Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.456490 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerDied","Data":"ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93"} Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.457538 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-vpzjk" event={"ID":"01bb709f-7c93-45e7-b900-802665c0225f","Type":"ContainerStarted","Data":"682e5d147d28fd0a49a14821c981d2edda88646f4847a2eab128c899b4592ca5"} Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.458939 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerID="c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6" exitCode=0 Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.458986 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerDied","Data":"c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6"} Dec 03 17:11:29 crc kubenswrapper[4758]: I1203 17:11:29.461146 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" event={"ID":"fa19d1a0-7212-4b46-8877-9e4e09f1f775","Type":"ContainerStarted","Data":"d2741b2862271b8bd10287cd7c6eaa4f2a7ba789948247f1b686f3aea9093a5b"} Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.179034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.199948 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/56cd1dfc-f881-441e-b231-d2783ddcf0f9-memberlist\") pod \"speaker-zb4zw\" (UID: \"56cd1dfc-f881-441e-b231-d2783ddcf0f9\") " pod="metallb-system/speaker-zb4zw" Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.337223 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zb4zw" Dec 03 17:11:30 crc kubenswrapper[4758]: W1203 17:11:30.364770 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56cd1dfc_f881_441e_b231_d2783ddcf0f9.slice/crio-65e9729718d610d8838b84b04fa820c96d74f8bb5a9c04e418e4fd8f3066fae5 WatchSource:0}: Error finding container 65e9729718d610d8838b84b04fa820c96d74f8bb5a9c04e418e4fd8f3066fae5: Status 404 returned error can't find the container with id 65e9729718d610d8838b84b04fa820c96d74f8bb5a9c04e418e4fd8f3066fae5 Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.474435 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-vpzjk" event={"ID":"01bb709f-7c93-45e7-b900-802665c0225f","Type":"ContainerStarted","Data":"038e9bd4487f0b556333cf82ffef905b57b2a56b7c5d21bd98d0de472430a475"} Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.475731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zb4zw" event={"ID":"56cd1dfc-f881-441e-b231-d2783ddcf0f9","Type":"ContainerStarted","Data":"65e9729718d610d8838b84b04fa820c96d74f8bb5a9c04e418e4fd8f3066fae5"} Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.479786 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"bebc827a892d476b238fcba8ae3ecb3b63efec28bf5a0fcaefcb9cdc94ffd13e"} Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.897268 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sxzdg"] Dec 03 17:11:30 crc kubenswrapper[4758]: I1203 17:11:30.897561 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sxzdg" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="registry-server" containerID="cri-o://c13229032f889eb33baff056451eee10715e89c0af754666c8df0dac41cb79b0" gracePeriod=2 Dec 03 17:11:31 crc kubenswrapper[4758]: I1203 17:11:31.496982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-vpzjk" event={"ID":"01bb709f-7c93-45e7-b900-802665c0225f","Type":"ContainerStarted","Data":"f7975a55eef805d77b057f95fdcce4d6ee917ed843431a27eefcae95b4a23a3b"} Dec 03 17:11:31 crc kubenswrapper[4758]: I1203 17:11:31.497174 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:31 crc kubenswrapper[4758]: I1203 17:11:31.509206 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zb4zw" event={"ID":"56cd1dfc-f881-441e-b231-d2783ddcf0f9","Type":"ContainerStarted","Data":"756c87906856206be653a1076f6636666c2089896c414d7a6540d8cc6bb2b64a"} Dec 03 17:11:31 crc kubenswrapper[4758]: I1203 17:11:31.524466 4758 generic.go:334] "Generic (PLEG): container finished" podID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerID="c13229032f889eb33baff056451eee10715e89c0af754666c8df0dac41cb79b0" exitCode=0 Dec 03 17:11:31 crc kubenswrapper[4758]: I1203 17:11:31.524535 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxzdg" event={"ID":"b3b8ac8b-1223-4e5b-8435-6f543ede9341","Type":"ContainerDied","Data":"c13229032f889eb33baff056451eee10715e89c0af754666c8df0dac41cb79b0"} Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.537745 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerStarted","Data":"4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2"} Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.541936 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zb4zw" event={"ID":"56cd1dfc-f881-441e-b231-d2783ddcf0f9","Type":"ContainerStarted","Data":"ec3bb02f8ff1e863397afb3477b6bcc77e72b2cf0b565a56a4b46494145c4f25"} Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.542422 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zb4zw" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.566469 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-vpzjk" podStartSLOduration=4.566442253 podStartE2EDuration="4.566442253s" podCreationTimestamp="2025-12-03 17:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:11:31.533805725 +0000 UTC m=+946.735182586" watchObservedRunningTime="2025-12-03 17:11:32.566442253 +0000 UTC m=+947.767819114" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.591936 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zb4zw" podStartSLOduration=4.591917702 podStartE2EDuration="4.591917702s" podCreationTimestamp="2025-12-03 17:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:11:32.587163545 +0000 UTC m=+947.788540406" watchObservedRunningTime="2025-12-03 17:11:32.591917702 +0000 UTC m=+947.793294563" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.724023 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.837938 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d868d\" (UniqueName: \"kubernetes.io/projected/b3b8ac8b-1223-4e5b-8435-6f543ede9341-kube-api-access-d868d\") pod \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.838616 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-catalog-content\") pod \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.838708 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-utilities\") pod \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\" (UID: \"b3b8ac8b-1223-4e5b-8435-6f543ede9341\") " Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.839927 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-utilities" (OuterVolumeSpecName: "utilities") pod "b3b8ac8b-1223-4e5b-8435-6f543ede9341" (UID: "b3b8ac8b-1223-4e5b-8435-6f543ede9341"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.845772 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b8ac8b-1223-4e5b-8435-6f543ede9341-kube-api-access-d868d" (OuterVolumeSpecName: "kube-api-access-d868d") pod "b3b8ac8b-1223-4e5b-8435-6f543ede9341" (UID: "b3b8ac8b-1223-4e5b-8435-6f543ede9341"). InnerVolumeSpecName "kube-api-access-d868d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.891973 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3b8ac8b-1223-4e5b-8435-6f543ede9341" (UID: "b3b8ac8b-1223-4e5b-8435-6f543ede9341"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.941187 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d868d\" (UniqueName: \"kubernetes.io/projected/b3b8ac8b-1223-4e5b-8435-6f543ede9341-kube-api-access-d868d\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.941226 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:32 crc kubenswrapper[4758]: I1203 17:11:32.941236 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b8ac8b-1223-4e5b-8435-6f543ede9341-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.552695 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxzdg" event={"ID":"b3b8ac8b-1223-4e5b-8435-6f543ede9341","Type":"ContainerDied","Data":"d3948678b6587e0762114a4fb3730b10e1dde8b157de04c17dfb2e8b26dede1b"} Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.552725 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxzdg" Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.552758 4758 scope.go:117] "RemoveContainer" containerID="c13229032f889eb33baff056451eee10715e89c0af754666c8df0dac41cb79b0" Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.560186 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerStarted","Data":"06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8"} Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.565828 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerID="4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2" exitCode=0 Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.565955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerDied","Data":"4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2"} Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.579948 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sxzdg"] Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.580006 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sxzdg"] Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.595063 4758 scope.go:117] "RemoveContainer" containerID="dafed731dfaa6e30e88ea170cfaf3ed2159ed88d88c0ddfe13f52584d07bde06" Dec 03 17:11:33 crc kubenswrapper[4758]: I1203 17:11:33.623253 4758 scope.go:117] "RemoveContainer" containerID="78dab2dd331fae43b01c253dabbff4a93d822bc60126dc00b04eaae42aab76d9" Dec 03 17:11:34 crc kubenswrapper[4758]: I1203 17:11:34.578152 4758 generic.go:334] "Generic (PLEG): container finished" podID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerID="06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8" exitCode=0 Dec 03 17:11:34 crc kubenswrapper[4758]: I1203 17:11:34.578213 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerDied","Data":"06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8"} Dec 03 17:11:34 crc kubenswrapper[4758]: I1203 17:11:34.584117 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerStarted","Data":"5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79"} Dec 03 17:11:34 crc kubenswrapper[4758]: I1203 17:11:34.643814 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ws8lt" podStartSLOduration=5.992797438 podStartE2EDuration="10.643788862s" podCreationTimestamp="2025-12-03 17:11:24 +0000 UTC" firstStartedPulling="2025-12-03 17:11:29.460739358 +0000 UTC m=+944.662116219" lastFinishedPulling="2025-12-03 17:11:34.111730782 +0000 UTC m=+949.313107643" observedRunningTime="2025-12-03 17:11:34.641092534 +0000 UTC m=+949.842469395" watchObservedRunningTime="2025-12-03 17:11:34.643788862 +0000 UTC m=+949.845165723" Dec 03 17:11:35 crc kubenswrapper[4758]: I1203 17:11:35.128856 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" path="/var/lib/kubelet/pods/b3b8ac8b-1223-4e5b-8435-6f543ede9341/volumes" Dec 03 17:11:36 crc kubenswrapper[4758]: I1203 17:11:36.617734 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerStarted","Data":"2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835"} Dec 03 17:11:36 crc kubenswrapper[4758]: I1203 17:11:36.648651 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cftwk" podStartSLOduration=4.291186955 podStartE2EDuration="9.648613535s" podCreationTimestamp="2025-12-03 17:11:27 +0000 UTC" firstStartedPulling="2025-12-03 17:11:30.547101333 +0000 UTC m=+945.748478204" lastFinishedPulling="2025-12-03 17:11:35.904527923 +0000 UTC m=+951.105904784" observedRunningTime="2025-12-03 17:11:36.646740931 +0000 UTC m=+951.848117792" watchObservedRunningTime="2025-12-03 17:11:36.648613535 +0000 UTC m=+951.849990406" Dec 03 17:11:37 crc kubenswrapper[4758]: I1203 17:11:37.843460 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:37 crc kubenswrapper[4758]: I1203 17:11:37.846374 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:38 crc kubenswrapper[4758]: I1203 17:11:38.933056 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-cftwk" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="registry-server" probeResult="failure" output=< Dec 03 17:11:38 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 17:11:38 crc kubenswrapper[4758]: > Dec 03 17:11:41 crc kubenswrapper[4758]: I1203 17:11:41.395203 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:11:41 crc kubenswrapper[4758]: I1203 17:11:41.395658 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:11:42 crc kubenswrapper[4758]: I1203 17:11:42.676038 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" event={"ID":"fa19d1a0-7212-4b46-8877-9e4e09f1f775","Type":"ContainerStarted","Data":"a344e94f3120c8b7185bea094ec37fdbc0019c3c787fb3d676fff9d591bffd33"} Dec 03 17:11:42 crc kubenswrapper[4758]: I1203 17:11:42.676185 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:42 crc kubenswrapper[4758]: I1203 17:11:42.678926 4758 generic.go:334] "Generic (PLEG): container finished" podID="272d2f0a-3217-4e67-82f3-43dc7aa334f5" containerID="8aad4bf85b9786ff7e1f7d6d2e0e031bddc95d0d6e1c2a3e630007ff0f021646" exitCode=0 Dec 03 17:11:42 crc kubenswrapper[4758]: I1203 17:11:42.679095 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerDied","Data":"8aad4bf85b9786ff7e1f7d6d2e0e031bddc95d0d6e1c2a3e630007ff0f021646"} Dec 03 17:11:42 crc kubenswrapper[4758]: I1203 17:11:42.703881 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" podStartSLOduration=1.731669898 podStartE2EDuration="14.70384263s" podCreationTimestamp="2025-12-03 17:11:28 +0000 UTC" firstStartedPulling="2025-12-03 17:11:29.01775479 +0000 UTC m=+944.219131641" lastFinishedPulling="2025-12-03 17:11:41.989927482 +0000 UTC m=+957.191304373" observedRunningTime="2025-12-03 17:11:42.694501092 +0000 UTC m=+957.895877953" watchObservedRunningTime="2025-12-03 17:11:42.70384263 +0000 UTC m=+957.905219501" Dec 03 17:11:43 crc kubenswrapper[4758]: I1203 17:11:43.690940 4758 generic.go:334] "Generic (PLEG): container finished" podID="272d2f0a-3217-4e67-82f3-43dc7aa334f5" containerID="967c13e1112671fc425bcae90ad894a4f1bfe46b50271f35d70dcf2ba3ae9eff" exitCode=0 Dec 03 17:11:43 crc kubenswrapper[4758]: I1203 17:11:43.691055 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerDied","Data":"967c13e1112671fc425bcae90ad894a4f1bfe46b50271f35d70dcf2ba3ae9eff"} Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.446513 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.446936 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.490936 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.702158 4758 generic.go:334] "Generic (PLEG): container finished" podID="272d2f0a-3217-4e67-82f3-43dc7aa334f5" containerID="5a7f0bad83ffd7a748858cdd0365a939e4b8fbd108f19206ff175fbd4411c493" exitCode=0 Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.702219 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerDied","Data":"5a7f0bad83ffd7a748858cdd0365a939e4b8fbd108f19206ff175fbd4411c493"} Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.782562 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:44 crc kubenswrapper[4758]: I1203 17:11:44.836281 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws8lt"] Dec 03 17:11:45 crc kubenswrapper[4758]: I1203 17:11:45.712611 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"d3f44702b0c532c3948c7e2e18e66fde4ecd87f49aca58c0b8bf2d2818df9d6e"} Dec 03 17:11:46 crc kubenswrapper[4758]: I1203 17:11:46.727916 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"976f9b1fc7cbe66ce039e12260d28e4b53b96afd97c32a8ac1a74eadb2d35263"} Dec 03 17:11:46 crc kubenswrapper[4758]: I1203 17:11:46.728386 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"dcd528cce088abcb00647b132a83ae5936bdb9236eeae1e334a346573bb9c55e"} Dec 03 17:11:46 crc kubenswrapper[4758]: I1203 17:11:46.728159 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ws8lt" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="registry-server" containerID="cri-o://5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79" gracePeriod=2 Dec 03 17:11:46 crc kubenswrapper[4758]: I1203 17:11:46.728403 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"ec4b8231e86f7edac2914c5b8524b92d0175ef0d954c0ea602882b69422c29e5"} Dec 03 17:11:46 crc kubenswrapper[4758]: I1203 17:11:46.728571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"1e88d4d82af0ddf06a06d0272c5f978f1afc9d0e1dc059aecbba8236d26a1b4d"} Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.128981 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.230602 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-catalog-content\") pod \"5b5a2278-6648-4466-8d43-c747ca02d55a\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.230739 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-utilities\") pod \"5b5a2278-6648-4466-8d43-c747ca02d55a\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.230820 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9p49\" (UniqueName: \"kubernetes.io/projected/5b5a2278-6648-4466-8d43-c747ca02d55a-kube-api-access-d9p49\") pod \"5b5a2278-6648-4466-8d43-c747ca02d55a\" (UID: \"5b5a2278-6648-4466-8d43-c747ca02d55a\") " Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.233022 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-utilities" (OuterVolumeSpecName: "utilities") pod "5b5a2278-6648-4466-8d43-c747ca02d55a" (UID: "5b5a2278-6648-4466-8d43-c747ca02d55a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.237032 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b5a2278-6648-4466-8d43-c747ca02d55a-kube-api-access-d9p49" (OuterVolumeSpecName: "kube-api-access-d9p49") pod "5b5a2278-6648-4466-8d43-c747ca02d55a" (UID: "5b5a2278-6648-4466-8d43-c747ca02d55a"). InnerVolumeSpecName "kube-api-access-d9p49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.249319 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b5a2278-6648-4466-8d43-c747ca02d55a" (UID: "5b5a2278-6648-4466-8d43-c747ca02d55a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.333588 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.333618 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b5a2278-6648-4466-8d43-c747ca02d55a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.333630 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9p49\" (UniqueName: \"kubernetes.io/projected/5b5a2278-6648-4466-8d43-c747ca02d55a-kube-api-access-d9p49\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.740248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxqxh" event={"ID":"272d2f0a-3217-4e67-82f3-43dc7aa334f5","Type":"ContainerStarted","Data":"ad711f2d87c8a9c6a43d263f3bbfc843ab776b54d321abdb5c1666f8e6931ecc"} Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.740857 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.744622 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws8lt" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.744633 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerID="5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79" exitCode=0 Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.744792 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerDied","Data":"5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79"} Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.744836 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws8lt" event={"ID":"5b5a2278-6648-4466-8d43-c747ca02d55a","Type":"ContainerDied","Data":"4ba878c2d1a080d75a922d04b08af26c1e26b36883cd406ff3579aa55d20dd93"} Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.744860 4758 scope.go:117] "RemoveContainer" containerID="5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.772912 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-gxqxh" podStartSLOduration=7.707651852 podStartE2EDuration="19.772883457s" podCreationTimestamp="2025-12-03 17:11:28 +0000 UTC" firstStartedPulling="2025-12-03 17:11:29.94120783 +0000 UTC m=+945.142584691" lastFinishedPulling="2025-12-03 17:11:42.006439395 +0000 UTC m=+957.207816296" observedRunningTime="2025-12-03 17:11:47.766727051 +0000 UTC m=+962.968103922" watchObservedRunningTime="2025-12-03 17:11:47.772883457 +0000 UTC m=+962.974260318" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.774398 4758 scope.go:117] "RemoveContainer" containerID="4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.789147 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws8lt"] Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.794569 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws8lt"] Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.809401 4758 scope.go:117] "RemoveContainer" containerID="c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.828496 4758 scope.go:117] "RemoveContainer" containerID="5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79" Dec 03 17:11:47 crc kubenswrapper[4758]: E1203 17:11:47.828961 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79\": container with ID starting with 5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79 not found: ID does not exist" containerID="5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.829026 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79"} err="failed to get container status \"5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79\": rpc error: code = NotFound desc = could not find container \"5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79\": container with ID starting with 5387506bc359e976e9b3f4718f1c69b01c830116eff5aedd72f724476a2cea79 not found: ID does not exist" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.829063 4758 scope.go:117] "RemoveContainer" containerID="4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2" Dec 03 17:11:47 crc kubenswrapper[4758]: E1203 17:11:47.829370 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2\": container with ID starting with 4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2 not found: ID does not exist" containerID="4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.829431 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2"} err="failed to get container status \"4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2\": rpc error: code = NotFound desc = could not find container \"4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2\": container with ID starting with 4035248eb2f4a4e8c0432f5dd677984d4194b9520926ec1f055f99fbe2df2ff2 not found: ID does not exist" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.829451 4758 scope.go:117] "RemoveContainer" containerID="c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6" Dec 03 17:11:47 crc kubenswrapper[4758]: E1203 17:11:47.829736 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6\": container with ID starting with c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6 not found: ID does not exist" containerID="c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.829771 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6"} err="failed to get container status \"c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6\": rpc error: code = NotFound desc = could not find container \"c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6\": container with ID starting with c69c3884d13d5af19b9b41e7a483354a5a2934aa93fe1bb60006efc1b58dc1e6 not found: ID does not exist" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.891589 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:47 crc kubenswrapper[4758]: I1203 17:11:47.949985 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:48 crc kubenswrapper[4758]: I1203 17:11:48.867192 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-vpzjk" Dec 03 17:11:49 crc kubenswrapper[4758]: I1203 17:11:49.122861 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" path="/var/lib/kubelet/pods/5b5a2278-6648-4466-8d43-c747ca02d55a/volumes" Dec 03 17:11:49 crc kubenswrapper[4758]: I1203 17:11:49.230636 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:49 crc kubenswrapper[4758]: I1203 17:11:49.272984 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.133909 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cftwk"] Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.134751 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cftwk" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="registry-server" containerID="cri-o://2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835" gracePeriod=2 Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.342217 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zb4zw" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.582798 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.597984 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-utilities\") pod \"de9ffab2-2a7e-4146-96bf-1a97167257d6\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.598093 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks4ws\" (UniqueName: \"kubernetes.io/projected/de9ffab2-2a7e-4146-96bf-1a97167257d6-kube-api-access-ks4ws\") pod \"de9ffab2-2a7e-4146-96bf-1a97167257d6\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.598194 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-catalog-content\") pod \"de9ffab2-2a7e-4146-96bf-1a97167257d6\" (UID: \"de9ffab2-2a7e-4146-96bf-1a97167257d6\") " Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.599192 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-utilities" (OuterVolumeSpecName: "utilities") pod "de9ffab2-2a7e-4146-96bf-1a97167257d6" (UID: "de9ffab2-2a7e-4146-96bf-1a97167257d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.620987 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9ffab2-2a7e-4146-96bf-1a97167257d6-kube-api-access-ks4ws" (OuterVolumeSpecName: "kube-api-access-ks4ws") pod "de9ffab2-2a7e-4146-96bf-1a97167257d6" (UID: "de9ffab2-2a7e-4146-96bf-1a97167257d6"). InnerVolumeSpecName "kube-api-access-ks4ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.672468 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de9ffab2-2a7e-4146-96bf-1a97167257d6" (UID: "de9ffab2-2a7e-4146-96bf-1a97167257d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.699031 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.699070 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks4ws\" (UniqueName: \"kubernetes.io/projected/de9ffab2-2a7e-4146-96bf-1a97167257d6-kube-api-access-ks4ws\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.699086 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9ffab2-2a7e-4146-96bf-1a97167257d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.769726 4758 generic.go:334] "Generic (PLEG): container finished" podID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerID="2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835" exitCode=0 Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.769807 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cftwk" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.769810 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerDied","Data":"2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835"} Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.770499 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cftwk" event={"ID":"de9ffab2-2a7e-4146-96bf-1a97167257d6","Type":"ContainerDied","Data":"a6764f269385ee4b10a43b5243f636c732d6afea93998e29ed7930c3d827e31c"} Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.770532 4758 scope.go:117] "RemoveContainer" containerID="2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.788821 4758 scope.go:117] "RemoveContainer" containerID="06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.814562 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cftwk"] Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.820122 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cftwk"] Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.825373 4758 scope.go:117] "RemoveContainer" containerID="ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.850222 4758 scope.go:117] "RemoveContainer" containerID="2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835" Dec 03 17:11:50 crc kubenswrapper[4758]: E1203 17:11:50.851315 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835\": container with ID starting with 2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835 not found: ID does not exist" containerID="2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.851364 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835"} err="failed to get container status \"2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835\": rpc error: code = NotFound desc = could not find container \"2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835\": container with ID starting with 2069e7d7d47ac3ba1995c6ea62835fc9302baa0bddd2187f6d1107fcafcb8835 not found: ID does not exist" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.851416 4758 scope.go:117] "RemoveContainer" containerID="06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8" Dec 03 17:11:50 crc kubenswrapper[4758]: E1203 17:11:50.852091 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8\": container with ID starting with 06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8 not found: ID does not exist" containerID="06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.852121 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8"} err="failed to get container status \"06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8\": rpc error: code = NotFound desc = could not find container \"06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8\": container with ID starting with 06122e73d7a8539f09f9f4a0a7286f95caa6669fe79cff5097d5b26dc0a041f8 not found: ID does not exist" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.852144 4758 scope.go:117] "RemoveContainer" containerID="ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93" Dec 03 17:11:50 crc kubenswrapper[4758]: E1203 17:11:50.852633 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93\": container with ID starting with ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93 not found: ID does not exist" containerID="ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93" Dec 03 17:11:50 crc kubenswrapper[4758]: I1203 17:11:50.852667 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93"} err="failed to get container status \"ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93\": rpc error: code = NotFound desc = could not find container \"ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93\": container with ID starting with ae258a275196bd68062d4b405e1af8824d1490f7a90afab462672ae0816c3e93 not found: ID does not exist" Dec 03 17:11:51 crc kubenswrapper[4758]: I1203 17:11:51.123221 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" path="/var/lib/kubelet/pods/de9ffab2-2a7e-4146-96bf-1a97167257d6/volumes" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187327 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8"] Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187568 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="extract-utilities" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187581 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="extract-utilities" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187594 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="extract-utilities" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187600 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="extract-utilities" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187611 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="extract-content" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187618 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="extract-content" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187629 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187635 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187644 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="extract-content" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187651 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="extract-content" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187658 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187664 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187700 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="extract-content" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187708 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="extract-content" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187719 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187726 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: E1203 17:11:52.187741 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="extract-utilities" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187747 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="extract-utilities" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187850 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b8ac8b-1223-4e5b-8435-6f543ede9341" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187859 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b5a2278-6648-4466-8d43-c747ca02d55a" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.187867 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9ffab2-2a7e-4146-96bf-1a97167257d6" containerName="registry-server" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.188866 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.202250 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.205853 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8"] Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.231465 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.231539 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn8b9\" (UniqueName: \"kubernetes.io/projected/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-kube-api-access-gn8b9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.231863 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.333153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.333243 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.333267 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn8b9\" (UniqueName: \"kubernetes.io/projected/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-kube-api-access-gn8b9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.333863 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.333986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.355572 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn8b9\" (UniqueName: \"kubernetes.io/projected/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-kube-api-access-gn8b9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:52 crc kubenswrapper[4758]: I1203 17:11:52.509258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:11:53 crc kubenswrapper[4758]: I1203 17:11:53.093750 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8"] Dec 03 17:11:53 crc kubenswrapper[4758]: I1203 17:11:53.801149 4758 generic.go:334] "Generic (PLEG): container finished" podID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerID="edf6b35e95cb603acaba352907c847383534c907f82bc226d0a58480ec1a607c" exitCode=0 Dec 03 17:11:53 crc kubenswrapper[4758]: I1203 17:11:53.801259 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" event={"ID":"24b483db-d9a3-4e07-ac9b-8f5c2b908bad","Type":"ContainerDied","Data":"edf6b35e95cb603acaba352907c847383534c907f82bc226d0a58480ec1a607c"} Dec 03 17:11:53 crc kubenswrapper[4758]: I1203 17:11:53.801608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" event={"ID":"24b483db-d9a3-4e07-ac9b-8f5c2b908bad","Type":"ContainerStarted","Data":"be23c72326de7bf5b842211a1491fb9ad8a10d1a61f1ef5fed35fbded0d0e543"} Dec 03 17:11:58 crc kubenswrapper[4758]: I1203 17:11:58.668287 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nqt6f" Dec 03 17:11:59 crc kubenswrapper[4758]: I1203 17:11:59.246332 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-gxqxh" Dec 03 17:12:00 crc kubenswrapper[4758]: I1203 17:12:00.854946 4758 generic.go:334] "Generic (PLEG): container finished" podID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerID="ee1b5c2c04d9a846bef75319afa0c4e5ab09ff6f1e7df3c32835019c742fa743" exitCode=0 Dec 03 17:12:00 crc kubenswrapper[4758]: I1203 17:12:00.855047 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" event={"ID":"24b483db-d9a3-4e07-ac9b-8f5c2b908bad","Type":"ContainerDied","Data":"ee1b5c2c04d9a846bef75319afa0c4e5ab09ff6f1e7df3c32835019c742fa743"} Dec 03 17:12:01 crc kubenswrapper[4758]: I1203 17:12:01.866090 4758 generic.go:334] "Generic (PLEG): container finished" podID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerID="b0870d1dd37cce8769e14c4fdb9bb3a4960af45f519603f3c4eb3cb3f4d849e6" exitCode=0 Dec 03 17:12:01 crc kubenswrapper[4758]: I1203 17:12:01.866201 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" event={"ID":"24b483db-d9a3-4e07-ac9b-8f5c2b908bad","Type":"ContainerDied","Data":"b0870d1dd37cce8769e14c4fdb9bb3a4960af45f519603f3c4eb3cb3f4d849e6"} Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.147755 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.324101 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-bundle\") pod \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.324221 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-util\") pod \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.324322 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn8b9\" (UniqueName: \"kubernetes.io/projected/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-kube-api-access-gn8b9\") pod \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\" (UID: \"24b483db-d9a3-4e07-ac9b-8f5c2b908bad\") " Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.325624 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-bundle" (OuterVolumeSpecName: "bundle") pod "24b483db-d9a3-4e07-ac9b-8f5c2b908bad" (UID: "24b483db-d9a3-4e07-ac9b-8f5c2b908bad"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.332717 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-kube-api-access-gn8b9" (OuterVolumeSpecName: "kube-api-access-gn8b9") pod "24b483db-d9a3-4e07-ac9b-8f5c2b908bad" (UID: "24b483db-d9a3-4e07-ac9b-8f5c2b908bad"). InnerVolumeSpecName "kube-api-access-gn8b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.334568 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-util" (OuterVolumeSpecName: "util") pod "24b483db-d9a3-4e07-ac9b-8f5c2b908bad" (UID: "24b483db-d9a3-4e07-ac9b-8f5c2b908bad"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.427258 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.427314 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn8b9\" (UniqueName: \"kubernetes.io/projected/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-kube-api-access-gn8b9\") on node \"crc\" DevicePath \"\"" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.427340 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/24b483db-d9a3-4e07-ac9b-8f5c2b908bad-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.883432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" event={"ID":"24b483db-d9a3-4e07-ac9b-8f5c2b908bad","Type":"ContainerDied","Data":"be23c72326de7bf5b842211a1491fb9ad8a10d1a61f1ef5fed35fbded0d0e543"} Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.883505 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be23c72326de7bf5b842211a1491fb9ad8a10d1a61f1ef5fed35fbded0d0e543" Dec 03 17:12:03 crc kubenswrapper[4758]: I1203 17:12:03.883613 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.824396 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n"] Dec 03 17:12:09 crc kubenswrapper[4758]: E1203 17:12:09.825519 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="pull" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.825539 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="pull" Dec 03 17:12:09 crc kubenswrapper[4758]: E1203 17:12:09.825553 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="extract" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.825560 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="extract" Dec 03 17:12:09 crc kubenswrapper[4758]: E1203 17:12:09.825573 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="util" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.825579 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="util" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.825715 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="24b483db-d9a3-4e07-ac9b-8f5c2b908bad" containerName="extract" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.826169 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.829337 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.829581 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.843036 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-pkqjh" Dec 03 17:12:09 crc kubenswrapper[4758]: I1203 17:12:09.855812 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n"] Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.019863 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c4259896-fda8-4b62-9321-93bad4a6b66e-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-h586n\" (UID: \"c4259896-fda8-4b62-9321-93bad4a6b66e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.019942 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbxlc\" (UniqueName: \"kubernetes.io/projected/c4259896-fda8-4b62-9321-93bad4a6b66e-kube-api-access-fbxlc\") pod \"cert-manager-operator-controller-manager-64cf6dff88-h586n\" (UID: \"c4259896-fda8-4b62-9321-93bad4a6b66e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.121992 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c4259896-fda8-4b62-9321-93bad4a6b66e-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-h586n\" (UID: \"c4259896-fda8-4b62-9321-93bad4a6b66e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.122192 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbxlc\" (UniqueName: \"kubernetes.io/projected/c4259896-fda8-4b62-9321-93bad4a6b66e-kube-api-access-fbxlc\") pod \"cert-manager-operator-controller-manager-64cf6dff88-h586n\" (UID: \"c4259896-fda8-4b62-9321-93bad4a6b66e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.122715 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c4259896-fda8-4b62-9321-93bad4a6b66e-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-h586n\" (UID: \"c4259896-fda8-4b62-9321-93bad4a6b66e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.146730 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbxlc\" (UniqueName: \"kubernetes.io/projected/c4259896-fda8-4b62-9321-93bad4a6b66e-kube-api-access-fbxlc\") pod \"cert-manager-operator-controller-manager-64cf6dff88-h586n\" (UID: \"c4259896-fda8-4b62-9321-93bad4a6b66e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.446771 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.787494 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n"] Dec 03 17:12:10 crc kubenswrapper[4758]: I1203 17:12:10.933947 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" event={"ID":"c4259896-fda8-4b62-9321-93bad4a6b66e","Type":"ContainerStarted","Data":"b30e776337d68a75c0c9313fa737d5499b51bae8d04f8f7f4ef0f9b73c9492ee"} Dec 03 17:12:11 crc kubenswrapper[4758]: I1203 17:12:11.394442 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:12:11 crc kubenswrapper[4758]: I1203 17:12:11.394501 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:12:11 crc kubenswrapper[4758]: I1203 17:12:11.394546 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:12:11 crc kubenswrapper[4758]: I1203 17:12:11.395206 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8050d5b5272390ff909cae77aca4e99c2b09448e05fbe4bc6797c541002e26b0"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:12:11 crc kubenswrapper[4758]: I1203 17:12:11.395265 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://8050d5b5272390ff909cae77aca4e99c2b09448e05fbe4bc6797c541002e26b0" gracePeriod=600 Dec 03 17:12:13 crc kubenswrapper[4758]: I1203 17:12:13.968603 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="8050d5b5272390ff909cae77aca4e99c2b09448e05fbe4bc6797c541002e26b0" exitCode=0 Dec 03 17:12:13 crc kubenswrapper[4758]: I1203 17:12:13.968749 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"8050d5b5272390ff909cae77aca4e99c2b09448e05fbe4bc6797c541002e26b0"} Dec 03 17:12:13 crc kubenswrapper[4758]: I1203 17:12:13.969537 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"74575219fc052d90d1263cdeb0cdc0186ad14da295a8ec4c660422997cca91c3"} Dec 03 17:12:13 crc kubenswrapper[4758]: I1203 17:12:13.969566 4758 scope.go:117] "RemoveContainer" containerID="174745d2d4c87b79a75b724a8f3fd0dcd39be458fd52fd0a6f156eafe510a247" Dec 03 17:12:23 crc kubenswrapper[4758]: I1203 17:12:23.084172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" event={"ID":"c4259896-fda8-4b62-9321-93bad4a6b66e","Type":"ContainerStarted","Data":"644e4a7b9119080bbcf24e1ee602fc45c67cf2f01ab35eedbe1465382ed6e0ed"} Dec 03 17:12:23 crc kubenswrapper[4758]: I1203 17:12:23.110773 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-h586n" podStartSLOduration=2.962571972 podStartE2EDuration="14.110744743s" podCreationTimestamp="2025-12-03 17:12:09 +0000 UTC" firstStartedPulling="2025-12-03 17:12:10.814628694 +0000 UTC m=+986.016005555" lastFinishedPulling="2025-12-03 17:12:21.962801475 +0000 UTC m=+997.164178326" observedRunningTime="2025-12-03 17:12:23.108585814 +0000 UTC m=+998.309962685" watchObservedRunningTime="2025-12-03 17:12:23.110744743 +0000 UTC m=+998.312121604" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.642460 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2"] Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.644137 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.646329 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-9gmwz" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.646777 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.647891 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.659999 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2"] Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.728299 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvbmv\" (UniqueName: \"kubernetes.io/projected/cefd882c-e7fc-4a9e-bcb4-2d349b184f7c-kube-api-access-hvbmv\") pod \"cert-manager-cainjector-855d9ccff4-wlvn2\" (UID: \"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.728368 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cefd882c-e7fc-4a9e-bcb4-2d349b184f7c-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-wlvn2\" (UID: \"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.830511 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvbmv\" (UniqueName: \"kubernetes.io/projected/cefd882c-e7fc-4a9e-bcb4-2d349b184f7c-kube-api-access-hvbmv\") pod \"cert-manager-cainjector-855d9ccff4-wlvn2\" (UID: \"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.830580 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cefd882c-e7fc-4a9e-bcb4-2d349b184f7c-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-wlvn2\" (UID: \"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.852401 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvbmv\" (UniqueName: \"kubernetes.io/projected/cefd882c-e7fc-4a9e-bcb4-2d349b184f7c-kube-api-access-hvbmv\") pod \"cert-manager-cainjector-855d9ccff4-wlvn2\" (UID: \"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.853268 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cefd882c-e7fc-4a9e-bcb4-2d349b184f7c-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-wlvn2\" (UID: \"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:26 crc kubenswrapper[4758]: I1203 17:12:26.965657 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" Dec 03 17:12:27 crc kubenswrapper[4758]: I1203 17:12:27.234736 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2"] Dec 03 17:12:28 crc kubenswrapper[4758]: I1203 17:12:28.128562 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" event={"ID":"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c","Type":"ContainerStarted","Data":"0644881c192bf537246d8629af687495f816c913ae7d92e3c81df2b486a93742"} Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.591284 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-rf2cm"] Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.592986 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.599869 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-rf2cm"] Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.600566 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-n8mdz" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.620596 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tdzx\" (UniqueName: \"kubernetes.io/projected/bf8ec7f2-ab50-4adc-b3f2-f19511690ddd-kube-api-access-8tdzx\") pod \"cert-manager-webhook-f4fb5df64-rf2cm\" (UID: \"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.620668 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf8ec7f2-ab50-4adc-b3f2-f19511690ddd-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-rf2cm\" (UID: \"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.721981 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tdzx\" (UniqueName: \"kubernetes.io/projected/bf8ec7f2-ab50-4adc-b3f2-f19511690ddd-kube-api-access-8tdzx\") pod \"cert-manager-webhook-f4fb5df64-rf2cm\" (UID: \"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.722066 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf8ec7f2-ab50-4adc-b3f2-f19511690ddd-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-rf2cm\" (UID: \"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.748077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf8ec7f2-ab50-4adc-b3f2-f19511690ddd-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-rf2cm\" (UID: \"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.764229 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tdzx\" (UniqueName: \"kubernetes.io/projected/bf8ec7f2-ab50-4adc-b3f2-f19511690ddd-kube-api-access-8tdzx\") pod \"cert-manager-webhook-f4fb5df64-rf2cm\" (UID: \"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:31 crc kubenswrapper[4758]: I1203 17:12:31.921219 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:36 crc kubenswrapper[4758]: I1203 17:12:36.913193 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-rf2cm"] Dec 03 17:12:37 crc kubenswrapper[4758]: I1203 17:12:37.220910 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" event={"ID":"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd","Type":"ContainerStarted","Data":"f198292d6edf2076869beb932dc41ba41c24474762b672d3e309e12e7d0b69a9"} Dec 03 17:12:37 crc kubenswrapper[4758]: I1203 17:12:37.220982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" event={"ID":"bf8ec7f2-ab50-4adc-b3f2-f19511690ddd","Type":"ContainerStarted","Data":"49d200f1d0c705c7eb96124f5e75096ae3bef5d86e645137c619dce844096e07"} Dec 03 17:12:37 crc kubenswrapper[4758]: I1203 17:12:37.221101 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:37 crc kubenswrapper[4758]: I1203 17:12:37.223010 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" event={"ID":"cefd882c-e7fc-4a9e-bcb4-2d349b184f7c","Type":"ContainerStarted","Data":"5677f2d0fac7cd306357322efd04ef3c1a79397bd41a3df17b8c260518cbfce4"} Dec 03 17:12:37 crc kubenswrapper[4758]: I1203 17:12:37.242813 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" podStartSLOduration=6.24278997 podStartE2EDuration="6.24278997s" podCreationTimestamp="2025-12-03 17:12:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:12:37.23909356 +0000 UTC m=+1012.440470751" watchObservedRunningTime="2025-12-03 17:12:37.24278997 +0000 UTC m=+1012.444166831" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.605330 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wlvn2" podStartSLOduration=10.289326029 podStartE2EDuration="19.605281594s" podCreationTimestamp="2025-12-03 17:12:26 +0000 UTC" firstStartedPulling="2025-12-03 17:12:27.244007387 +0000 UTC m=+1002.445384248" lastFinishedPulling="2025-12-03 17:12:36.559962952 +0000 UTC m=+1011.761339813" observedRunningTime="2025-12-03 17:12:37.261244874 +0000 UTC m=+1012.462621725" watchObservedRunningTime="2025-12-03 17:12:45.605281594 +0000 UTC m=+1020.806658455" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.615184 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-q85ck"] Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.617763 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.620757 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-trx4l" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.626144 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-q85ck"] Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.651641 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbzth\" (UniqueName: \"kubernetes.io/projected/8cb254c1-10c6-4f10-b201-85f3b5b66b03-kube-api-access-nbzth\") pod \"cert-manager-86cb77c54b-q85ck\" (UID: \"8cb254c1-10c6-4f10-b201-85f3b5b66b03\") " pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.651805 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cb254c1-10c6-4f10-b201-85f3b5b66b03-bound-sa-token\") pod \"cert-manager-86cb77c54b-q85ck\" (UID: \"8cb254c1-10c6-4f10-b201-85f3b5b66b03\") " pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.753978 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbzth\" (UniqueName: \"kubernetes.io/projected/8cb254c1-10c6-4f10-b201-85f3b5b66b03-kube-api-access-nbzth\") pod \"cert-manager-86cb77c54b-q85ck\" (UID: \"8cb254c1-10c6-4f10-b201-85f3b5b66b03\") " pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.754065 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cb254c1-10c6-4f10-b201-85f3b5b66b03-bound-sa-token\") pod \"cert-manager-86cb77c54b-q85ck\" (UID: \"8cb254c1-10c6-4f10-b201-85f3b5b66b03\") " pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.777206 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbzth\" (UniqueName: \"kubernetes.io/projected/8cb254c1-10c6-4f10-b201-85f3b5b66b03-kube-api-access-nbzth\") pod \"cert-manager-86cb77c54b-q85ck\" (UID: \"8cb254c1-10c6-4f10-b201-85f3b5b66b03\") " pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.777453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cb254c1-10c6-4f10-b201-85f3b5b66b03-bound-sa-token\") pod \"cert-manager-86cb77c54b-q85ck\" (UID: \"8cb254c1-10c6-4f10-b201-85f3b5b66b03\") " pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:45 crc kubenswrapper[4758]: I1203 17:12:45.940257 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-q85ck" Dec 03 17:12:46 crc kubenswrapper[4758]: I1203 17:12:46.349786 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-q85ck"] Dec 03 17:12:46 crc kubenswrapper[4758]: I1203 17:12:46.925127 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-rf2cm" Dec 03 17:12:47 crc kubenswrapper[4758]: I1203 17:12:47.319952 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-q85ck" event={"ID":"8cb254c1-10c6-4f10-b201-85f3b5b66b03","Type":"ContainerStarted","Data":"6bc9d143412b114014bd96cae0a240afaa73b25852b45dbebeded8b4814f538b"} Dec 03 17:12:47 crc kubenswrapper[4758]: I1203 17:12:47.320018 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-q85ck" event={"ID":"8cb254c1-10c6-4f10-b201-85f3b5b66b03","Type":"ContainerStarted","Data":"4650e288ae333e35bad31238bd43abdc053d4357a919b38f1cfe85d2d3d74f24"} Dec 03 17:12:47 crc kubenswrapper[4758]: I1203 17:12:47.340304 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-q85ck" podStartSLOduration=2.340273703 podStartE2EDuration="2.340273703s" podCreationTimestamp="2025-12-03 17:12:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:12:47.338188175 +0000 UTC m=+1022.539565036" watchObservedRunningTime="2025-12-03 17:12:47.340273703 +0000 UTC m=+1022.541650564" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.247122 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-m7xkk"] Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.248803 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.253567 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fj52n" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.253742 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.253586 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.277051 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m7xkk"] Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.331184 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzd42\" (UniqueName: \"kubernetes.io/projected/a82d3b5a-67f5-4634-8f9f-3562a3c39e27-kube-api-access-zzd42\") pod \"openstack-operator-index-m7xkk\" (UID: \"a82d3b5a-67f5-4634-8f9f-3562a3c39e27\") " pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.433086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzd42\" (UniqueName: \"kubernetes.io/projected/a82d3b5a-67f5-4634-8f9f-3562a3c39e27-kube-api-access-zzd42\") pod \"openstack-operator-index-m7xkk\" (UID: \"a82d3b5a-67f5-4634-8f9f-3562a3c39e27\") " pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.459188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzd42\" (UniqueName: \"kubernetes.io/projected/a82d3b5a-67f5-4634-8f9f-3562a3c39e27-kube-api-access-zzd42\") pod \"openstack-operator-index-m7xkk\" (UID: \"a82d3b5a-67f5-4634-8f9f-3562a3c39e27\") " pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.594115 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:12:50 crc kubenswrapper[4758]: I1203 17:12:50.834334 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m7xkk"] Dec 03 17:12:50 crc kubenswrapper[4758]: W1203 17:12:50.840579 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda82d3b5a_67f5_4634_8f9f_3562a3c39e27.slice/crio-4eba39c4333414d0dc5d15143758c4a8cf8fee4986cfee94f9e0114621016952 WatchSource:0}: Error finding container 4eba39c4333414d0dc5d15143758c4a8cf8fee4986cfee94f9e0114621016952: Status 404 returned error can't find the container with id 4eba39c4333414d0dc5d15143758c4a8cf8fee4986cfee94f9e0114621016952 Dec 03 17:12:51 crc kubenswrapper[4758]: I1203 17:12:51.367111 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m7xkk" event={"ID":"a82d3b5a-67f5-4634-8f9f-3562a3c39e27","Type":"ContainerStarted","Data":"4eba39c4333414d0dc5d15143758c4a8cf8fee4986cfee94f9e0114621016952"} Dec 03 17:12:53 crc kubenswrapper[4758]: I1203 17:12:53.621324 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-m7xkk"] Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.228483 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-48xhl"] Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.229636 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.239672 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-48xhl"] Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.290928 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff2f6\" (UniqueName: \"kubernetes.io/projected/ee9c8f83-7b21-4742-9e08-90c355c3c0fe-kube-api-access-ff2f6\") pod \"openstack-operator-index-48xhl\" (UID: \"ee9c8f83-7b21-4742-9e08-90c355c3c0fe\") " pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.393129 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff2f6\" (UniqueName: \"kubernetes.io/projected/ee9c8f83-7b21-4742-9e08-90c355c3c0fe-kube-api-access-ff2f6\") pod \"openstack-operator-index-48xhl\" (UID: \"ee9c8f83-7b21-4742-9e08-90c355c3c0fe\") " pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.435057 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff2f6\" (UniqueName: \"kubernetes.io/projected/ee9c8f83-7b21-4742-9e08-90c355c3c0fe-kube-api-access-ff2f6\") pod \"openstack-operator-index-48xhl\" (UID: \"ee9c8f83-7b21-4742-9e08-90c355c3c0fe\") " pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.550828 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:12:54 crc kubenswrapper[4758]: I1203 17:12:54.761602 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-48xhl"] Dec 03 17:12:54 crc kubenswrapper[4758]: W1203 17:12:54.770226 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9c8f83_7b21_4742_9e08_90c355c3c0fe.slice/crio-12ebdcae06ff760dc545c4ec902e85b53c5454a7b6f0c8cc6db6ca6efc323fbb WatchSource:0}: Error finding container 12ebdcae06ff760dc545c4ec902e85b53c5454a7b6f0c8cc6db6ca6efc323fbb: Status 404 returned error can't find the container with id 12ebdcae06ff760dc545c4ec902e85b53c5454a7b6f0c8cc6db6ca6efc323fbb Dec 03 17:12:55 crc kubenswrapper[4758]: I1203 17:12:55.405926 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-48xhl" event={"ID":"ee9c8f83-7b21-4742-9e08-90c355c3c0fe","Type":"ContainerStarted","Data":"12ebdcae06ff760dc545c4ec902e85b53c5454a7b6f0c8cc6db6ca6efc323fbb"} Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.439414 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-48xhl" event={"ID":"ee9c8f83-7b21-4742-9e08-90c355c3c0fe","Type":"ContainerStarted","Data":"c0cc3762509e7c58d2b268b66cce9f467e41140bc95afa0c6e8d2ebf7da3fbbd"} Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.442851 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m7xkk" event={"ID":"a82d3b5a-67f5-4634-8f9f-3562a3c39e27","Type":"ContainerStarted","Data":"568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26"} Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.443062 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-m7xkk" podUID="a82d3b5a-67f5-4634-8f9f-3562a3c39e27" containerName="registry-server" containerID="cri-o://568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26" gracePeriod=2 Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.462146 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-48xhl" podStartSLOduration=1.604344883 podStartE2EDuration="5.4621096s" podCreationTimestamp="2025-12-03 17:12:54 +0000 UTC" firstStartedPulling="2025-12-03 17:12:54.775818464 +0000 UTC m=+1029.977195325" lastFinishedPulling="2025-12-03 17:12:58.633583181 +0000 UTC m=+1033.834960042" observedRunningTime="2025-12-03 17:12:59.456023914 +0000 UTC m=+1034.657400775" watchObservedRunningTime="2025-12-03 17:12:59.4621096 +0000 UTC m=+1034.663486461" Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.489663 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-m7xkk" podStartSLOduration=1.702627348 podStartE2EDuration="9.489632732s" podCreationTimestamp="2025-12-03 17:12:50 +0000 UTC" firstStartedPulling="2025-12-03 17:12:50.844345955 +0000 UTC m=+1026.045722816" lastFinishedPulling="2025-12-03 17:12:58.631351339 +0000 UTC m=+1033.832728200" observedRunningTime="2025-12-03 17:12:59.486430065 +0000 UTC m=+1034.687806926" watchObservedRunningTime="2025-12-03 17:12:59.489632732 +0000 UTC m=+1034.691009593" Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.979347 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.985623 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzd42\" (UniqueName: \"kubernetes.io/projected/a82d3b5a-67f5-4634-8f9f-3562a3c39e27-kube-api-access-zzd42\") pod \"a82d3b5a-67f5-4634-8f9f-3562a3c39e27\" (UID: \"a82d3b5a-67f5-4634-8f9f-3562a3c39e27\") " Dec 03 17:12:59 crc kubenswrapper[4758]: I1203 17:12:59.993637 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82d3b5a-67f5-4634-8f9f-3562a3c39e27-kube-api-access-zzd42" (OuterVolumeSpecName: "kube-api-access-zzd42") pod "a82d3b5a-67f5-4634-8f9f-3562a3c39e27" (UID: "a82d3b5a-67f5-4634-8f9f-3562a3c39e27"). InnerVolumeSpecName "kube-api-access-zzd42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.086788 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzd42\" (UniqueName: \"kubernetes.io/projected/a82d3b5a-67f5-4634-8f9f-3562a3c39e27-kube-api-access-zzd42\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.451850 4758 generic.go:334] "Generic (PLEG): container finished" podID="a82d3b5a-67f5-4634-8f9f-3562a3c39e27" containerID="568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26" exitCode=0 Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.451919 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m7xkk" Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.451955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m7xkk" event={"ID":"a82d3b5a-67f5-4634-8f9f-3562a3c39e27","Type":"ContainerDied","Data":"568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26"} Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.452078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m7xkk" event={"ID":"a82d3b5a-67f5-4634-8f9f-3562a3c39e27","Type":"ContainerDied","Data":"4eba39c4333414d0dc5d15143758c4a8cf8fee4986cfee94f9e0114621016952"} Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.452115 4758 scope.go:117] "RemoveContainer" containerID="568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26" Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.476527 4758 scope.go:117] "RemoveContainer" containerID="568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26" Dec 03 17:13:00 crc kubenswrapper[4758]: E1203 17:13:00.477654 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26\": container with ID starting with 568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26 not found: ID does not exist" containerID="568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26" Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.477740 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26"} err="failed to get container status \"568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26\": rpc error: code = NotFound desc = could not find container \"568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26\": container with ID starting with 568215d3e6a035ebdbb409869e9a02fcdc5e03aa4e5fc0c34bd00ded7af24a26 not found: ID does not exist" Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.489601 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-m7xkk"] Dec 03 17:13:00 crc kubenswrapper[4758]: I1203 17:13:00.495272 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-m7xkk"] Dec 03 17:13:01 crc kubenswrapper[4758]: I1203 17:13:01.125134 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a82d3b5a-67f5-4634-8f9f-3562a3c39e27" path="/var/lib/kubelet/pods/a82d3b5a-67f5-4634-8f9f-3562a3c39e27/volumes" Dec 03 17:13:04 crc kubenswrapper[4758]: I1203 17:13:04.551770 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:13:04 crc kubenswrapper[4758]: I1203 17:13:04.552284 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:13:04 crc kubenswrapper[4758]: I1203 17:13:04.724166 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:13:05 crc kubenswrapper[4758]: I1203 17:13:05.630755 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-48xhl" Dec 03 17:13:12 crc kubenswrapper[4758]: I1203 17:13:12.846918 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4"] Dec 03 17:13:12 crc kubenswrapper[4758]: E1203 17:13:12.847827 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82d3b5a-67f5-4634-8f9f-3562a3c39e27" containerName="registry-server" Dec 03 17:13:12 crc kubenswrapper[4758]: I1203 17:13:12.847844 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82d3b5a-67f5-4634-8f9f-3562a3c39e27" containerName="registry-server" Dec 03 17:13:12 crc kubenswrapper[4758]: I1203 17:13:12.847997 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82d3b5a-67f5-4634-8f9f-3562a3c39e27" containerName="registry-server" Dec 03 17:13:12 crc kubenswrapper[4758]: I1203 17:13:12.848938 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:12 crc kubenswrapper[4758]: I1203 17:13:12.851597 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xv5tb" Dec 03 17:13:12 crc kubenswrapper[4758]: I1203 17:13:12.898612 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4"] Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.046158 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbbt\" (UniqueName: \"kubernetes.io/projected/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-kube-api-access-blbbt\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.046261 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-bundle\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.046336 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-util\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.148176 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-util\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.148281 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbbt\" (UniqueName: \"kubernetes.io/projected/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-kube-api-access-blbbt\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.148330 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-bundle\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.148862 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-util\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.148924 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-bundle\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.171103 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbbt\" (UniqueName: \"kubernetes.io/projected/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-kube-api-access-blbbt\") pod \"432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:13 crc kubenswrapper[4758]: I1203 17:13:13.469947 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:14 crc kubenswrapper[4758]: I1203 17:13:14.158481 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4"] Dec 03 17:13:14 crc kubenswrapper[4758]: I1203 17:13:14.572726 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" event={"ID":"43b518b0-93c4-45bc-8b3c-a997dd2a02ee","Type":"ContainerStarted","Data":"01b1a30f26efca3647d13427e4f66ce84b0fcc0dd68515576928d2de1f002658"} Dec 03 17:13:16 crc kubenswrapper[4758]: I1203 17:13:16.644084 4758 generic.go:334] "Generic (PLEG): container finished" podID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerID="b59b9f306bdd6059a9b7766184e250aad4dc1b80e1f88bededfd24a3e70d72b4" exitCode=0 Dec 03 17:13:16 crc kubenswrapper[4758]: I1203 17:13:16.644191 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" event={"ID":"43b518b0-93c4-45bc-8b3c-a997dd2a02ee","Type":"ContainerDied","Data":"b59b9f306bdd6059a9b7766184e250aad4dc1b80e1f88bededfd24a3e70d72b4"} Dec 03 17:13:18 crc kubenswrapper[4758]: I1203 17:13:18.667246 4758 generic.go:334] "Generic (PLEG): container finished" podID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerID="4ca17c6f653377b53091ed48fdf3b89b651297328f9ec857003e828ba0f58da8" exitCode=0 Dec 03 17:13:18 crc kubenswrapper[4758]: I1203 17:13:18.667587 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" event={"ID":"43b518b0-93c4-45bc-8b3c-a997dd2a02ee","Type":"ContainerDied","Data":"4ca17c6f653377b53091ed48fdf3b89b651297328f9ec857003e828ba0f58da8"} Dec 03 17:13:19 crc kubenswrapper[4758]: I1203 17:13:19.684463 4758 generic.go:334] "Generic (PLEG): container finished" podID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerID="3d789733530bc0ce112ae1533f2e0b745994ad8d07436631fff2c3f24722cd12" exitCode=0 Dec 03 17:13:19 crc kubenswrapper[4758]: I1203 17:13:19.684591 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" event={"ID":"43b518b0-93c4-45bc-8b3c-a997dd2a02ee","Type":"ContainerDied","Data":"3d789733530bc0ce112ae1533f2e0b745994ad8d07436631fff2c3f24722cd12"} Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.918057 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.944873 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-util\") pod \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.945208 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blbbt\" (UniqueName: \"kubernetes.io/projected/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-kube-api-access-blbbt\") pod \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.945331 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-bundle\") pod \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\" (UID: \"43b518b0-93c4-45bc-8b3c-a997dd2a02ee\") " Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.946697 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-bundle" (OuterVolumeSpecName: "bundle") pod "43b518b0-93c4-45bc-8b3c-a997dd2a02ee" (UID: "43b518b0-93c4-45bc-8b3c-a997dd2a02ee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.951870 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-kube-api-access-blbbt" (OuterVolumeSpecName: "kube-api-access-blbbt") pod "43b518b0-93c4-45bc-8b3c-a997dd2a02ee" (UID: "43b518b0-93c4-45bc-8b3c-a997dd2a02ee"). InnerVolumeSpecName "kube-api-access-blbbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:20 crc kubenswrapper[4758]: I1203 17:13:20.960111 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-util" (OuterVolumeSpecName: "util") pod "43b518b0-93c4-45bc-8b3c-a997dd2a02ee" (UID: "43b518b0-93c4-45bc-8b3c-a997dd2a02ee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:21 crc kubenswrapper[4758]: I1203 17:13:21.047062 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:21 crc kubenswrapper[4758]: I1203 17:13:21.047110 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:21 crc kubenswrapper[4758]: I1203 17:13:21.047123 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blbbt\" (UniqueName: \"kubernetes.io/projected/43b518b0-93c4-45bc-8b3c-a997dd2a02ee-kube-api-access-blbbt\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:21 crc kubenswrapper[4758]: I1203 17:13:21.703834 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" event={"ID":"43b518b0-93c4-45bc-8b3c-a997dd2a02ee","Type":"ContainerDied","Data":"01b1a30f26efca3647d13427e4f66ce84b0fcc0dd68515576928d2de1f002658"} Dec 03 17:13:21 crc kubenswrapper[4758]: I1203 17:13:21.703896 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01b1a30f26efca3647d13427e4f66ce84b0fcc0dd68515576928d2de1f002658" Dec 03 17:13:21 crc kubenswrapper[4758]: I1203 17:13:21.704042 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.850233 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq"] Dec 03 17:13:24 crc kubenswrapper[4758]: E1203 17:13:24.851421 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="extract" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.851443 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="extract" Dec 03 17:13:24 crc kubenswrapper[4758]: E1203 17:13:24.851466 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="util" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.851474 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="util" Dec 03 17:13:24 crc kubenswrapper[4758]: E1203 17:13:24.851502 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="pull" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.851509 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="pull" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.851657 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="43b518b0-93c4-45bc-8b3c-a997dd2a02ee" containerName="extract" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.852365 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.854289 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-lcljw" Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.870914 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq"] Dec 03 17:13:24 crc kubenswrapper[4758]: I1203 17:13:24.915189 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2987t\" (UniqueName: \"kubernetes.io/projected/b2f53a34-0ea1-4d23-8b26-a077003f0f9c-kube-api-access-2987t\") pod \"openstack-operator-controller-operator-6b5fc879f-9f9cq\" (UID: \"b2f53a34-0ea1-4d23-8b26-a077003f0f9c\") " pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:13:25 crc kubenswrapper[4758]: I1203 17:13:25.017346 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2987t\" (UniqueName: \"kubernetes.io/projected/b2f53a34-0ea1-4d23-8b26-a077003f0f9c-kube-api-access-2987t\") pod \"openstack-operator-controller-operator-6b5fc879f-9f9cq\" (UID: \"b2f53a34-0ea1-4d23-8b26-a077003f0f9c\") " pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:13:25 crc kubenswrapper[4758]: I1203 17:13:25.036981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2987t\" (UniqueName: \"kubernetes.io/projected/b2f53a34-0ea1-4d23-8b26-a077003f0f9c-kube-api-access-2987t\") pod \"openstack-operator-controller-operator-6b5fc879f-9f9cq\" (UID: \"b2f53a34-0ea1-4d23-8b26-a077003f0f9c\") " pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:13:25 crc kubenswrapper[4758]: I1203 17:13:25.173913 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:13:25 crc kubenswrapper[4758]: I1203 17:13:25.645443 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq"] Dec 03 17:13:25 crc kubenswrapper[4758]: I1203 17:13:25.737515 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" event={"ID":"b2f53a34-0ea1-4d23-8b26-a077003f0f9c","Type":"ContainerStarted","Data":"3bdd763e6fc4e2b61ae5a674ef2a0882b196d98086d2c30f2c0d8d76b8438946"} Dec 03 17:13:36 crc kubenswrapper[4758]: I1203 17:13:36.869271 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" event={"ID":"b2f53a34-0ea1-4d23-8b26-a077003f0f9c","Type":"ContainerStarted","Data":"3daccd7e735055717aac1806512c52b53e181437416b617fae0bae9c057a648c"} Dec 03 17:13:36 crc kubenswrapper[4758]: I1203 17:13:36.870104 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:13:36 crc kubenswrapper[4758]: I1203 17:13:36.909347 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" podStartSLOduration=2.215972204 podStartE2EDuration="12.909314388s" podCreationTimestamp="2025-12-03 17:13:24 +0000 UTC" firstStartedPulling="2025-12-03 17:13:25.657903243 +0000 UTC m=+1060.859280104" lastFinishedPulling="2025-12-03 17:13:36.351245427 +0000 UTC m=+1071.552622288" observedRunningTime="2025-12-03 17:13:36.899368367 +0000 UTC m=+1072.100745228" watchObservedRunningTime="2025-12-03 17:13:36.909314388 +0000 UTC m=+1072.110691249" Dec 03 17:13:45 crc kubenswrapper[4758]: I1203 17:13:45.177413 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6b5fc879f-9f9cq" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.166758 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.168643 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.170952 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wgj5j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.182134 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.183562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.192459 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-q4jvk" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.199630 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.215745 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.216983 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.218400 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v22pt\" (UniqueName: \"kubernetes.io/projected/de31e4e2-fe86-4543-82c5-da042e77b6ce-kube-api-access-v22pt\") pod \"barbican-operator-controller-manager-7d9dfd778-fnlpm\" (UID: \"de31e4e2-fe86-4543-82c5-da042e77b6ce\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.218455 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjzpf\" (UniqueName: \"kubernetes.io/projected/36504cff-c657-485e-9db7-788382c4a9e1-kube-api-access-zjzpf\") pod \"cinder-operator-controller-manager-859b6ccc6-j4sfr\" (UID: \"36504cff-c657-485e-9db7-788382c4a9e1\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.223446 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-jq2cd" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.240937 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.242243 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.244628 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sndfh" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.244623 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.248565 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.251662 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-wkg26" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.262806 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.270747 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.276047 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.331172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k68r6\" (UniqueName: \"kubernetes.io/projected/84cea5a5-c77f-4818-9c41-8d91a838f7fa-kube-api-access-k68r6\") pod \"glance-operator-controller-manager-77987cd8cd-72tmg\" (UID: \"84cea5a5-c77f-4818-9c41-8d91a838f7fa\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.331588 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v22pt\" (UniqueName: \"kubernetes.io/projected/de31e4e2-fe86-4543-82c5-da042e77b6ce-kube-api-access-v22pt\") pod \"barbican-operator-controller-manager-7d9dfd778-fnlpm\" (UID: \"de31e4e2-fe86-4543-82c5-da042e77b6ce\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.332033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjzpf\" (UniqueName: \"kubernetes.io/projected/36504cff-c657-485e-9db7-788382c4a9e1-kube-api-access-zjzpf\") pod \"cinder-operator-controller-manager-859b6ccc6-j4sfr\" (UID: \"36504cff-c657-485e-9db7-788382c4a9e1\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.381005 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.381577 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v22pt\" (UniqueName: \"kubernetes.io/projected/de31e4e2-fe86-4543-82c5-da042e77b6ce-kube-api-access-v22pt\") pod \"barbican-operator-controller-manager-7d9dfd778-fnlpm\" (UID: \"de31e4e2-fe86-4543-82c5-da042e77b6ce\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.392243 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.393604 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.401344 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nftjz" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.402347 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjzpf\" (UniqueName: \"kubernetes.io/projected/36504cff-c657-485e-9db7-788382c4a9e1-kube-api-access-zjzpf\") pod \"cinder-operator-controller-manager-859b6ccc6-j4sfr\" (UID: \"36504cff-c657-485e-9db7-788382c4a9e1\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.435810 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhnnt\" (UniqueName: \"kubernetes.io/projected/5477637e-ea68-454e-8cf6-51b491464daf-kube-api-access-hhnnt\") pod \"heat-operator-controller-manager-5f64f6f8bb-4972h\" (UID: \"5477637e-ea68-454e-8cf6-51b491464daf\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.435872 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24wcv\" (UniqueName: \"kubernetes.io/projected/99462bc7-4e24-40b1-8ce5-555c47c8305f-kube-api-access-24wcv\") pod \"designate-operator-controller-manager-78b4bc895b-8ll9j\" (UID: \"99462bc7-4e24-40b1-8ce5-555c47c8305f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.435944 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k68r6\" (UniqueName: \"kubernetes.io/projected/84cea5a5-c77f-4818-9c41-8d91a838f7fa-kube-api-access-k68r6\") pod \"glance-operator-controller-manager-77987cd8cd-72tmg\" (UID: \"84cea5a5-c77f-4818-9c41-8d91a838f7fa\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.435985 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8t6j\" (UniqueName: \"kubernetes.io/projected/61019cc3-111f-4f59-8e43-64b5dfdd22c4-kube-api-access-j8t6j\") pod \"horizon-operator-controller-manager-68c6d99b8f-q8ddt\" (UID: \"61019cc3-111f-4f59-8e43-64b5dfdd22c4\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.448014 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.461609 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-7m28f"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.462991 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.466017 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9pvfw" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.468851 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.472341 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.473805 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.481276 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jrgqs" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.487713 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-7m28f"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.498929 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.502593 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.510952 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.512310 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.515160 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.516366 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-h5hwj" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.541633 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.543085 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k68r6\" (UniqueName: \"kubernetes.io/projected/84cea5a5-c77f-4818-9c41-8d91a838f7fa-kube-api-access-k68r6\") pod \"glance-operator-controller-manager-77987cd8cd-72tmg\" (UID: \"84cea5a5-c77f-4818-9c41-8d91a838f7fa\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.546160 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.549755 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.551739 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-xqcq5" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.567654 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhnnt\" (UniqueName: \"kubernetes.io/projected/5477637e-ea68-454e-8cf6-51b491464daf-kube-api-access-hhnnt\") pod \"heat-operator-controller-manager-5f64f6f8bb-4972h\" (UID: \"5477637e-ea68-454e-8cf6-51b491464daf\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.567760 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjgrt\" (UniqueName: \"kubernetes.io/projected/144cd5a6-c5c5-456d-b28d-27c3a3fb8898-kube-api-access-tjgrt\") pod \"ironic-operator-controller-manager-6c548fd776-5djcl\" (UID: \"144cd5a6-c5c5-456d-b28d-27c3a3fb8898\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.567807 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt9b5\" (UniqueName: \"kubernetes.io/projected/4ecb6824-b536-4432-8e37-92dc55964815-kube-api-access-jt9b5\") pod \"keystone-operator-controller-manager-7765d96ddf-r5r8j\" (UID: \"4ecb6824-b536-4432-8e37-92dc55964815\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.567836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24wcv\" (UniqueName: \"kubernetes.io/projected/99462bc7-4e24-40b1-8ce5-555c47c8305f-kube-api-access-24wcv\") pod \"designate-operator-controller-manager-78b4bc895b-8ll9j\" (UID: \"99462bc7-4e24-40b1-8ce5-555c47c8305f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.567880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.567932 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t86p\" (UniqueName: \"kubernetes.io/projected/1911d89e-c35c-4104-b0c1-265f770a731d-kube-api-access-8t86p\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.568016 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8t6j\" (UniqueName: \"kubernetes.io/projected/61019cc3-111f-4f59-8e43-64b5dfdd22c4-kube-api-access-j8t6j\") pod \"horizon-operator-controller-manager-68c6d99b8f-q8ddt\" (UID: \"61019cc3-111f-4f59-8e43-64b5dfdd22c4\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.579499 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.597058 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.619780 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.621153 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.625594 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24wcv\" (UniqueName: \"kubernetes.io/projected/99462bc7-4e24-40b1-8ce5-555c47c8305f-kube-api-access-24wcv\") pod \"designate-operator-controller-manager-78b4bc895b-8ll9j\" (UID: \"99462bc7-4e24-40b1-8ce5-555c47c8305f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.626456 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8t6j\" (UniqueName: \"kubernetes.io/projected/61019cc3-111f-4f59-8e43-64b5dfdd22c4-kube-api-access-j8t6j\") pod \"horizon-operator-controller-manager-68c6d99b8f-q8ddt\" (UID: \"61019cc3-111f-4f59-8e43-64b5dfdd22c4\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.626976 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhnnt\" (UniqueName: \"kubernetes.io/projected/5477637e-ea68-454e-8cf6-51b491464daf-kube-api-access-hhnnt\") pod \"heat-operator-controller-manager-5f64f6f8bb-4972h\" (UID: \"5477637e-ea68-454e-8cf6-51b491464daf\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.627328 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-8lgpn" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.648798 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.663819 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.665374 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.667940 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-8zgc8" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.675542 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.675622 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzhpn\" (UniqueName: \"kubernetes.io/projected/74b430d9-3316-4157-ab4f-264f7797c7e9-kube-api-access-jzhpn\") pod \"mariadb-operator-controller-manager-56bbcc9d85-ndmwx\" (UID: \"74b430d9-3316-4157-ab4f-264f7797c7e9\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.675660 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t86p\" (UniqueName: \"kubernetes.io/projected/1911d89e-c35c-4104-b0c1-265f770a731d-kube-api-access-8t86p\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.675716 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxpdx\" (UniqueName: \"kubernetes.io/projected/f6893abc-6551-4bc5-bcc5-18dd74da025b-kube-api-access-sxpdx\") pod \"manila-operator-controller-manager-7c79b5df47-qcpt6\" (UID: \"f6893abc-6551-4bc5-bcc5-18dd74da025b\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.675778 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgrt\" (UniqueName: \"kubernetes.io/projected/144cd5a6-c5c5-456d-b28d-27c3a3fb8898-kube-api-access-tjgrt\") pod \"ironic-operator-controller-manager-6c548fd776-5djcl\" (UID: \"144cd5a6-c5c5-456d-b28d-27c3a3fb8898\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.675807 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt9b5\" (UniqueName: \"kubernetes.io/projected/4ecb6824-b536-4432-8e37-92dc55964815-kube-api-access-jt9b5\") pod \"keystone-operator-controller-manager-7765d96ddf-r5r8j\" (UID: \"4ecb6824-b536-4432-8e37-92dc55964815\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:14:12 crc kubenswrapper[4758]: E1203 17:14:12.676280 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:12 crc kubenswrapper[4758]: E1203 17:14:12.676366 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert podName:1911d89e-c35c-4104-b0c1-265f770a731d nodeName:}" failed. No retries permitted until 2025-12-03 17:14:13.176340874 +0000 UTC m=+1108.377717735 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert") pod "infra-operator-controller-manager-57548d458d-7m28f" (UID: "1911d89e-c35c-4104-b0c1-265f770a731d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.698022 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.699357 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.705012 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.712583 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.715787 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.724278 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-g6bs6" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.732212 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.732400 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qm4d7" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.748893 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.755018 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgrt\" (UniqueName: \"kubernetes.io/projected/144cd5a6-c5c5-456d-b28d-27c3a3fb8898-kube-api-access-tjgrt\") pod \"ironic-operator-controller-manager-6c548fd776-5djcl\" (UID: \"144cd5a6-c5c5-456d-b28d-27c3a3fb8898\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.756811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt9b5\" (UniqueName: \"kubernetes.io/projected/4ecb6824-b536-4432-8e37-92dc55964815-kube-api-access-jt9b5\") pod \"keystone-operator-controller-manager-7765d96ddf-r5r8j\" (UID: \"4ecb6824-b536-4432-8e37-92dc55964815\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.757874 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.762555 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t86p\" (UniqueName: \"kubernetes.io/projected/1911d89e-c35c-4104-b0c1-265f770a731d-kube-api-access-8t86p\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.777700 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9wrt\" (UniqueName: \"kubernetes.io/projected/29c8457a-abba-43c0-b2ac-12c78bdeef4c-kube-api-access-g9wrt\") pod \"nova-operator-controller-manager-697bc559fc-c5wfm\" (UID: \"29c8457a-abba-43c0-b2ac-12c78bdeef4c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.777786 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxj28\" (UniqueName: \"kubernetes.io/projected/c624801b-66ee-4a15-866a-a65e7c69b2df-kube-api-access-xxj28\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5mgxc\" (UID: \"c624801b-66ee-4a15-866a-a65e7c69b2df\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.777813 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz6tw\" (UniqueName: \"kubernetes.io/projected/7806e6cc-eee5-4735-bae7-10ca5e4144d6-kube-api-access-lz6tw\") pod \"octavia-operator-controller-manager-998648c74-rp8pj\" (UID: \"7806e6cc-eee5-4735-bae7-10ca5e4144d6\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.777840 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzhpn\" (UniqueName: \"kubernetes.io/projected/74b430d9-3316-4157-ab4f-264f7797c7e9-kube-api-access-jzhpn\") pod \"mariadb-operator-controller-manager-56bbcc9d85-ndmwx\" (UID: \"74b430d9-3316-4157-ab4f-264f7797c7e9\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.777889 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxpdx\" (UniqueName: \"kubernetes.io/projected/f6893abc-6551-4bc5-bcc5-18dd74da025b-kube-api-access-sxpdx\") pod \"manila-operator-controller-manager-7c79b5df47-qcpt6\" (UID: \"f6893abc-6551-4bc5-bcc5-18dd74da025b\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.830014 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.831283 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.837442 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxpdx\" (UniqueName: \"kubernetes.io/projected/f6893abc-6551-4bc5-bcc5-18dd74da025b-kube-api-access-sxpdx\") pod \"manila-operator-controller-manager-7c79b5df47-qcpt6\" (UID: \"f6893abc-6551-4bc5-bcc5-18dd74da025b\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.838056 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.838700 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-stnfh" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.838856 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.880003 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9wrt\" (UniqueName: \"kubernetes.io/projected/29c8457a-abba-43c0-b2ac-12c78bdeef4c-kube-api-access-g9wrt\") pod \"nova-operator-controller-manager-697bc559fc-c5wfm\" (UID: \"29c8457a-abba-43c0-b2ac-12c78bdeef4c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.880080 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.880116 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxj28\" (UniqueName: \"kubernetes.io/projected/c624801b-66ee-4a15-866a-a65e7c69b2df-kube-api-access-xxj28\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5mgxc\" (UID: \"c624801b-66ee-4a15-866a-a65e7c69b2df\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.880140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz6tw\" (UniqueName: \"kubernetes.io/projected/7806e6cc-eee5-4735-bae7-10ca5e4144d6-kube-api-access-lz6tw\") pod \"octavia-operator-controller-manager-998648c74-rp8pj\" (UID: \"7806e6cc-eee5-4735-bae7-10ca5e4144d6\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.880236 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tggjs\" (UniqueName: \"kubernetes.io/projected/ff85f702-0ab2-490b-899b-e14bcd599eb4-kube-api-access-tggjs\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.901577 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzhpn\" (UniqueName: \"kubernetes.io/projected/74b430d9-3316-4157-ab4f-264f7797c7e9-kube-api-access-jzhpn\") pod \"mariadb-operator-controller-manager-56bbcc9d85-ndmwx\" (UID: \"74b430d9-3316-4157-ab4f-264f7797c7e9\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.912568 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.920183 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.951809 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j"] Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.982636 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9wrt\" (UniqueName: \"kubernetes.io/projected/29c8457a-abba-43c0-b2ac-12c78bdeef4c-kube-api-access-g9wrt\") pod \"nova-operator-controller-manager-697bc559fc-c5wfm\" (UID: \"29c8457a-abba-43c0-b2ac-12c78bdeef4c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.984474 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.985663 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxj28\" (UniqueName: \"kubernetes.io/projected/c624801b-66ee-4a15-866a-a65e7c69b2df-kube-api-access-xxj28\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-5mgxc\" (UID: \"c624801b-66ee-4a15-866a-a65e7c69b2df\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.985739 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tggjs\" (UniqueName: \"kubernetes.io/projected/ff85f702-0ab2-490b-899b-e14bcd599eb4-kube-api-access-tggjs\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.985839 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:12 crc kubenswrapper[4758]: E1203 17:14:12.986070 4758 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:12 crc kubenswrapper[4758]: E1203 17:14:12.986146 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert podName:ff85f702-0ab2-490b-899b-e14bcd599eb4 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:13.486126339 +0000 UTC m=+1108.687503200 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" (UID: "ff85f702-0ab2-490b-899b-e14bcd599eb4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:12 crc kubenswrapper[4758]: I1203 17:14:12.991173 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz6tw\" (UniqueName: \"kubernetes.io/projected/7806e6cc-eee5-4735-bae7-10ca5e4144d6-kube-api-access-lz6tw\") pod \"octavia-operator-controller-manager-998648c74-rp8pj\" (UID: \"7806e6cc-eee5-4735-bae7-10ca5e4144d6\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.019827 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.021199 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.026140 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-zr9lg" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.045060 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tggjs\" (UniqueName: \"kubernetes.io/projected/ff85f702-0ab2-490b-899b-e14bcd599eb4-kube-api-access-tggjs\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.045728 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.083948 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.109019 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l259\" (UniqueName: \"kubernetes.io/projected/ce961712-a4d4-4a93-96e6-0cf4fdc58eea-kube-api-access-7l259\") pod \"ovn-operator-controller-manager-b6456fdb6-p8mlv\" (UID: \"ce961712-a4d4-4a93-96e6-0cf4fdc58eea\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.114608 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.122578 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.137972 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.161546 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-54hks" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.173723 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.207179 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.207223 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.207246 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.208698 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.212174 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x5j2p" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.224494 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l259\" (UniqueName: \"kubernetes.io/projected/ce961712-a4d4-4a93-96e6-0cf4fdc58eea-kube-api-access-7l259\") pod \"ovn-operator-controller-manager-b6456fdb6-p8mlv\" (UID: \"ce961712-a4d4-4a93-96e6-0cf4fdc58eea\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.224553 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62tvv\" (UniqueName: \"kubernetes.io/projected/aae66ae2-507f-43f8-bac5-c842427bb780-kube-api-access-62tvv\") pod \"placement-operator-controller-manager-78f8948974-4pfjv\" (UID: \"aae66ae2-507f-43f8-bac5-c842427bb780\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.224640 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.224854 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.224930 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert podName:1911d89e-c35c-4104-b0c1-265f770a731d nodeName:}" failed. No retries permitted until 2025-12-03 17:14:14.224908306 +0000 UTC m=+1109.426285167 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert") pod "infra-operator-controller-manager-57548d458d-7m28f" (UID: "1911d89e-c35c-4104-b0c1-265f770a731d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.237611 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.266832 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.298477 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l259\" (UniqueName: \"kubernetes.io/projected/ce961712-a4d4-4a93-96e6-0cf4fdc58eea-kube-api-access-7l259\") pod \"ovn-operator-controller-manager-b6456fdb6-p8mlv\" (UID: \"ce961712-a4d4-4a93-96e6-0cf4fdc58eea\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.325215 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.328564 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssbtr\" (UniqueName: \"kubernetes.io/projected/eefa346f-ce5f-4445-8565-f6ad8250f049-kube-api-access-ssbtr\") pod \"swift-operator-controller-manager-5f8c65bbfc-2j99q\" (UID: \"eefa346f-ce5f-4445-8565-f6ad8250f049\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.328629 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62tvv\" (UniqueName: \"kubernetes.io/projected/aae66ae2-507f-43f8-bac5-c842427bb780-kube-api-access-62tvv\") pod \"placement-operator-controller-manager-78f8948974-4pfjv\" (UID: \"aae66ae2-507f-43f8-bac5-c842427bb780\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.335010 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-xrcwj" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.365093 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.378178 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.378656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62tvv\" (UniqueName: \"kubernetes.io/projected/aae66ae2-507f-43f8-bac5-c842427bb780-kube-api-access-62tvv\") pod \"placement-operator-controller-manager-78f8948974-4pfjv\" (UID: \"aae66ae2-507f-43f8-bac5-c842427bb780\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.424331 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-54w5p"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.432924 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-54w5p"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.436450 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh8s9\" (UniqueName: \"kubernetes.io/projected/61d9456a-70b6-4a3c-80fe-1ad9a65d0f44-kube-api-access-zh8s9\") pod \"telemetry-operator-controller-manager-76cc84c6bb-5x7rp\" (UID: \"61d9456a-70b6-4a3c-80fe-1ad9a65d0f44\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.436609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssbtr\" (UniqueName: \"kubernetes.io/projected/eefa346f-ce5f-4445-8565-f6ad8250f049-kube-api-access-ssbtr\") pod \"swift-operator-controller-manager-5f8c65bbfc-2j99q\" (UID: \"eefa346f-ce5f-4445-8565-f6ad8250f049\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.441154 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.452083 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-shdpd" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.463955 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.469606 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.472369 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-h8bnl" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.486777 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.493283 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssbtr\" (UniqueName: \"kubernetes.io/projected/eefa346f-ce5f-4445-8565-f6ad8250f049-kube-api-access-ssbtr\") pod \"swift-operator-controller-manager-5f8c65bbfc-2j99q\" (UID: \"eefa346f-ce5f-4445-8565-f6ad8250f049\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.503187 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.515486 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.516810 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.519193 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.520084 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.520300 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-z7htx" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.538913 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.538966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh8s9\" (UniqueName: \"kubernetes.io/projected/61d9456a-70b6-4a3c-80fe-1ad9a65d0f44-kube-api-access-zh8s9\") pod \"telemetry-operator-controller-manager-76cc84c6bb-5x7rp\" (UID: \"61d9456a-70b6-4a3c-80fe-1ad9a65d0f44\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.539011 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csf54\" (UniqueName: \"kubernetes.io/projected/68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1-kube-api-access-csf54\") pod \"watcher-operator-controller-manager-769dc69bc-6llwb\" (UID: \"68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.539044 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69hc9\" (UniqueName: \"kubernetes.io/projected/56b5e5d0-a007-4451-bfbb-87e6f7201367-kube-api-access-69hc9\") pod \"test-operator-controller-manager-5854674fcc-54w5p\" (UID: \"56b5e5d0-a007-4451-bfbb-87e6f7201367\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.539246 4758 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.539302 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert podName:ff85f702-0ab2-490b-899b-e14bcd599eb4 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:14.539280466 +0000 UTC m=+1109.740657317 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" (UID: "ff85f702-0ab2-490b-899b-e14bcd599eb4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.566093 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh8s9\" (UniqueName: \"kubernetes.io/projected/61d9456a-70b6-4a3c-80fe-1ad9a65d0f44-kube-api-access-zh8s9\") pod \"telemetry-operator-controller-manager-76cc84c6bb-5x7rp\" (UID: \"61d9456a-70b6-4a3c-80fe-1ad9a65d0f44\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.571063 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.614039 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.617476 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.627549 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.629211 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.635613 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-s7vq4" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.639729 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.640790 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf8bs\" (UniqueName: \"kubernetes.io/projected/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-kube-api-access-zf8bs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.640865 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.640901 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csf54\" (UniqueName: \"kubernetes.io/projected/68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1-kube-api-access-csf54\") pod \"watcher-operator-controller-manager-769dc69bc-6llwb\" (UID: \"68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.640930 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69hc9\" (UniqueName: \"kubernetes.io/projected/56b5e5d0-a007-4451-bfbb-87e6f7201367-kube-api-access-69hc9\") pod \"test-operator-controller-manager-5854674fcc-54w5p\" (UID: \"56b5e5d0-a007-4451-bfbb-87e6f7201367\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.640958 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.657983 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.664838 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.678708 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg"] Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.679607 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csf54\" (UniqueName: \"kubernetes.io/projected/68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1-kube-api-access-csf54\") pod \"watcher-operator-controller-manager-769dc69bc-6llwb\" (UID: \"68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.723105 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69hc9\" (UniqueName: \"kubernetes.io/projected/56b5e5d0-a007-4451-bfbb-87e6f7201367-kube-api-access-69hc9\") pod \"test-operator-controller-manager-5854674fcc-54w5p\" (UID: \"56b5e5d0-a007-4451-bfbb-87e6f7201367\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.761854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf8bs\" (UniqueName: \"kubernetes.io/projected/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-kube-api-access-zf8bs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.762019 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbwz4\" (UniqueName: \"kubernetes.io/projected/5a38f83d-d017-49fc-98bf-e08ae999428c-kube-api-access-jbwz4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4vmjv\" (UID: \"5a38f83d-d017-49fc-98bf-e08ae999428c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.762059 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.762108 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.763879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.764118 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.764252 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:14.264224745 +0000 UTC m=+1109.465601606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "webhook-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.765218 4758 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: E1203 17:14:13.765305 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:14.265271414 +0000 UTC m=+1109.466648275 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "metrics-server-cert" not found Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.829453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf8bs\" (UniqueName: \"kubernetes.io/projected/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-kube-api-access-zf8bs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.876641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbwz4\" (UniqueName: \"kubernetes.io/projected/5a38f83d-d017-49fc-98bf-e08ae999428c-kube-api-access-jbwz4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4vmjv\" (UID: \"5a38f83d-d017-49fc-98bf-e08ae999428c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.914899 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbwz4\" (UniqueName: \"kubernetes.io/projected/5a38f83d-d017-49fc-98bf-e08ae999428c-kube-api-access-jbwz4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4vmjv\" (UID: \"5a38f83d-d017-49fc-98bf-e08ae999428c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.924442 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" Dec 03 17:14:13 crc kubenswrapper[4758]: I1203 17:14:13.950500 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.298653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.299135 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.299223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.299394 4758 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.299466 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:15.299442593 +0000 UTC m=+1110.500819454 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "metrics-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.300012 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.300045 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:15.30003482 +0000 UTC m=+1110.501411681 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "webhook-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.300088 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.300106 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert podName:1911d89e-c35c-4104-b0c1-265f770a731d nodeName:}" failed. No retries permitted until 2025-12-03 17:14:16.300100422 +0000 UTC m=+1111.501477283 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert") pod "infra-operator-controller-manager-57548d458d-7m28f" (UID: "1911d89e-c35c-4104-b0c1-265f770a731d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.516833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" event={"ID":"36504cff-c657-485e-9db7-788382c4a9e1","Type":"ContainerStarted","Data":"9f17b1ff3e56095d7dd1f2acf7a5afd9d33e83c61bc302b3563fb4909b62fa18"} Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.529187 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" event={"ID":"84cea5a5-c77f-4818-9c41-8d91a838f7fa","Type":"ContainerStarted","Data":"7c14f1b6f988dba98e0bb2f55e4ced954ce4903bcefcf30d6dfd43b85356786e"} Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.563580 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm"] Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.587400 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt"] Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.608247 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.608358 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl"] Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.608612 4758 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: E1203 17:14:14.608704 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert podName:ff85f702-0ab2-490b-899b-e14bcd599eb4 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:16.608662793 +0000 UTC m=+1111.810039664 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" (UID: "ff85f702-0ab2-490b-899b-e14bcd599eb4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:14 crc kubenswrapper[4758]: I1203 17:14:14.621289 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.112525 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6"] Dec 03 17:14:15 crc kubenswrapper[4758]: W1203 17:14:15.125433 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6893abc_6551_4bc5_bcc5_18dd74da025b.slice/crio-18e51445930a5168b8b8f47b19b70ccf97fdcb141624942ddc8c9399a7fc680d WatchSource:0}: Error finding container 18e51445930a5168b8b8f47b19b70ccf97fdcb141624942ddc8c9399a7fc680d: Status 404 returned error can't find the container with id 18e51445930a5168b8b8f47b19b70ccf97fdcb141624942ddc8c9399a7fc680d Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.171761 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.188694 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.195989 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.203783 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.220847 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.236796 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv"] Dec 03 17:14:15 crc kubenswrapper[4758]: W1203 17:14:15.240751 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeefa346f_ce5f_4445_8565_f6ad8250f049.slice/crio-4b4e22f532be827a296a14aaf0dadceb027c0ccfcd74fe0dea77c554f14adbbd WatchSource:0}: Error finding container 4b4e22f532be827a296a14aaf0dadceb027c0ccfcd74fe0dea77c554f14adbbd: Status 404 returned error can't find the container with id 4b4e22f532be827a296a14aaf0dadceb027c0ccfcd74fe0dea77c554f14adbbd Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.242594 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-54w5p"] Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.244101 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-csf54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-6llwb_openstack-operators(68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.244151 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jbwz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-4vmjv_openstack-operators(5a38f83d-d017-49fc-98bf-e08ae999428c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.244163 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ssbtr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-2j99q_openstack-operators(eefa346f-ce5f-4445-8565-f6ad8250f049): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.245770 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" podUID="5a38f83d-d017-49fc-98bf-e08ae999428c" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.247169 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-csf54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-6llwb_openstack-operators(68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.248315 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" podUID="68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1" Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.250115 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.261107 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.266675 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.323991 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.324152 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.324286 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.324319 4758 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.324423 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:17.324387047 +0000 UTC m=+1112.525764068 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "webhook-server-cert" not found Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.324456 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:17.324443749 +0000 UTC m=+1112.525820620 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "metrics-server-cert" not found Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.542778 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" event={"ID":"f6893abc-6551-4bc5-bcc5-18dd74da025b","Type":"ContainerStarted","Data":"18e51445930a5168b8b8f47b19b70ccf97fdcb141624942ddc8c9399a7fc680d"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.547353 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" event={"ID":"ce961712-a4d4-4a93-96e6-0cf4fdc58eea","Type":"ContainerStarted","Data":"8668c3699fb121b701e71275c5424fb11a23b2dac32f430d5b09e993a480f37c"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.549460 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" event={"ID":"5a38f83d-d017-49fc-98bf-e08ae999428c","Type":"ContainerStarted","Data":"7c0a41c96612c4e9f7ae2a26d248c637e09cc7bab577727f9472e577cb669bbb"} Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.566200 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" podUID="5a38f83d-d017-49fc-98bf-e08ae999428c" Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.568230 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" event={"ID":"68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1","Type":"ContainerStarted","Data":"2d603ad605fd91faf745a39a0aefd90fcd19d47609623ed4119d8fc0c22efb0a"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.571049 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" event={"ID":"61d9456a-70b6-4a3c-80fe-1ad9a65d0f44","Type":"ContainerStarted","Data":"1e8e182eee6fc61c9dd46e0787e6d41f6364b47d47b82e9db20c7860f4848d84"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.572789 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" event={"ID":"c624801b-66ee-4a15-866a-a65e7c69b2df","Type":"ContainerStarted","Data":"5a5b86b94f4149db5c107594b52430bcca36c268f09e2d812fd8fc1051c4a1e3"} Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.574255 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" podUID="68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1" Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.575312 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" event={"ID":"144cd5a6-c5c5-456d-b28d-27c3a3fb8898","Type":"ContainerStarted","Data":"705deaab7e11530888f9216c83457a1ab21486241fdac7646d36548d0ed8c087"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.578381 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" event={"ID":"4ecb6824-b536-4432-8e37-92dc55964815","Type":"ContainerStarted","Data":"37cf08750e9df5cf8ef9e561888382caf69001bf5fcabb5bc7adbd011d69de62"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.583009 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" event={"ID":"5477637e-ea68-454e-8cf6-51b491464daf","Type":"ContainerStarted","Data":"ae002e24f29891c6242de94712c4c29fee620129ce709013389a7e299a5926ac"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.586004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" event={"ID":"eefa346f-ce5f-4445-8565-f6ad8250f049","Type":"ContainerStarted","Data":"4b4e22f532be827a296a14aaf0dadceb027c0ccfcd74fe0dea77c554f14adbbd"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.587657 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" event={"ID":"99462bc7-4e24-40b1-8ce5-555c47c8305f","Type":"ContainerStarted","Data":"288a4f74e480600e13cb37f1e327daa34e11405ece7fd06dc26ad24a531a62c3"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.597217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" event={"ID":"de31e4e2-fe86-4543-82c5-da042e77b6ce","Type":"ContainerStarted","Data":"95f42579a2c7862a24766f0fb40e7d32a22f8bb9130f517902aa27249ecc7291"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.598633 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" event={"ID":"29c8457a-abba-43c0-b2ac-12c78bdeef4c","Type":"ContainerStarted","Data":"d0d56644f5180ffdde60b22ebf9a927329886dfbff7939e22970254348c4dc08"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.602469 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" event={"ID":"56b5e5d0-a007-4451-bfbb-87e6f7201367","Type":"ContainerStarted","Data":"7842ceafb46b03fc995ae65caa3276fe0dc53e08c2185967bc56754aec227223"} Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.604500 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" event={"ID":"61019cc3-111f-4f59-8e43-64b5dfdd22c4","Type":"ContainerStarted","Data":"32207a66bb6428166c4670de407282d5e87809e18aecb0d1e39127318e56924b"} Dec 03 17:14:15 crc kubenswrapper[4758]: W1203 17:14:15.665953 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaae66ae2_507f_43f8_bac5_c842427bb780.slice/crio-1bd8ce248c58025b15cafd4be35ee1d5444365776863a232f671a18e56baa648 WatchSource:0}: Error finding container 1bd8ce248c58025b15cafd4be35ee1d5444365776863a232f671a18e56baa648: Status 404 returned error can't find the container with id 1bd8ce248c58025b15cafd4be35ee1d5444365776863a232f671a18e56baa648 Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.668915 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.685457 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv"] Dec 03 17:14:15 crc kubenswrapper[4758]: I1203 17:14:15.706305 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx"] Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.706519 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-62tvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4pfjv_openstack-operators(aae66ae2-507f-43f8-bac5-c842427bb780): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.709029 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-62tvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4pfjv_openstack-operators(aae66ae2-507f-43f8-bac5-c842427bb780): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:14:15 crc kubenswrapper[4758]: E1203 17:14:15.710403 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" podUID="aae66ae2-507f-43f8-bac5-c842427bb780" Dec 03 17:14:16 crc kubenswrapper[4758]: I1203 17:14:16.344805 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.345073 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.345194 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert podName:1911d89e-c35c-4104-b0c1-265f770a731d nodeName:}" failed. No retries permitted until 2025-12-03 17:14:20.345159947 +0000 UTC m=+1115.546536808 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert") pod "infra-operator-controller-manager-57548d458d-7m28f" (UID: "1911d89e-c35c-4104-b0c1-265f770a731d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:16 crc kubenswrapper[4758]: I1203 17:14:16.623584 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" event={"ID":"74b430d9-3316-4157-ab4f-264f7797c7e9","Type":"ContainerStarted","Data":"cc08dc59cb298808c11fc1acabbde2438eb718e5686bfbde8a9dec07b84efb23"} Dec 03 17:14:16 crc kubenswrapper[4758]: I1203 17:14:16.627865 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" event={"ID":"aae66ae2-507f-43f8-bac5-c842427bb780","Type":"ContainerStarted","Data":"1bd8ce248c58025b15cafd4be35ee1d5444365776863a232f671a18e56baa648"} Dec 03 17:14:16 crc kubenswrapper[4758]: I1203 17:14:16.629964 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" event={"ID":"7806e6cc-eee5-4735-bae7-10ca5e4144d6","Type":"ContainerStarted","Data":"894882fe61a1d936a9ed8c0e09c2d639771f9455411b97084fe35a17a6273ef9"} Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.632994 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" podUID="aae66ae2-507f-43f8-bac5-c842427bb780" Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.635260 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" podUID="5a38f83d-d017-49fc-98bf-e08ae999428c" Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.635855 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" podUID="68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1" Dec 03 17:14:16 crc kubenswrapper[4758]: I1203 17:14:16.649006 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.649284 4758 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:16 crc kubenswrapper[4758]: E1203 17:14:16.649422 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert podName:ff85f702-0ab2-490b-899b-e14bcd599eb4 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:20.64936486 +0000 UTC m=+1115.850741721 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" (UID: "ff85f702-0ab2-490b-899b-e14bcd599eb4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:17 crc kubenswrapper[4758]: I1203 17:14:17.375641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:17 crc kubenswrapper[4758]: I1203 17:14:17.376222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:17 crc kubenswrapper[4758]: E1203 17:14:17.376503 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:14:17 crc kubenswrapper[4758]: E1203 17:14:17.376586 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:21.376560637 +0000 UTC m=+1116.577937498 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "webhook-server-cert" not found Dec 03 17:14:17 crc kubenswrapper[4758]: E1203 17:14:17.377235 4758 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:14:17 crc kubenswrapper[4758]: E1203 17:14:17.377286 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:21.377271696 +0000 UTC m=+1116.578648577 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "metrics-server-cert" not found Dec 03 17:14:17 crc kubenswrapper[4758]: E1203 17:14:17.687355 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" podUID="aae66ae2-507f-43f8-bac5-c842427bb780" Dec 03 17:14:20 crc kubenswrapper[4758]: I1203 17:14:20.345201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:20 crc kubenswrapper[4758]: E1203 17:14:20.345719 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:20 crc kubenswrapper[4758]: E1203 17:14:20.345809 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert podName:1911d89e-c35c-4104-b0c1-265f770a731d nodeName:}" failed. No retries permitted until 2025-12-03 17:14:28.345782737 +0000 UTC m=+1123.547159598 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert") pod "infra-operator-controller-manager-57548d458d-7m28f" (UID: "1911d89e-c35c-4104-b0c1-265f770a731d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:14:20 crc kubenswrapper[4758]: I1203 17:14:20.667288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:20 crc kubenswrapper[4758]: E1203 17:14:20.667629 4758 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:20 crc kubenswrapper[4758]: E1203 17:14:20.667710 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert podName:ff85f702-0ab2-490b-899b-e14bcd599eb4 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:28.667688631 +0000 UTC m=+1123.869065492 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" (UID: "ff85f702-0ab2-490b-899b-e14bcd599eb4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:14:21 crc kubenswrapper[4758]: I1203 17:14:21.460639 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:21 crc kubenswrapper[4758]: I1203 17:14:21.460856 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:21 crc kubenswrapper[4758]: E1203 17:14:21.460918 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:14:21 crc kubenswrapper[4758]: E1203 17:14:21.461012 4758 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:14:21 crc kubenswrapper[4758]: E1203 17:14:21.461056 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:29.461017754 +0000 UTC m=+1124.662394605 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "webhook-server-cert" not found Dec 03 17:14:21 crc kubenswrapper[4758]: E1203 17:14:21.461099 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:29.461070906 +0000 UTC m=+1124.662447997 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "metrics-server-cert" not found Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.117270 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.415761 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.440806 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1911d89e-c35c-4104-b0c1-265f770a731d-cert\") pod \"infra-operator-controller-manager-57548d458d-7m28f\" (UID: \"1911d89e-c35c-4104-b0c1-265f770a731d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.689453 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.727406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.734978 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff85f702-0ab2-490b-899b-e14bcd599eb4-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j\" (UID: \"ff85f702-0ab2-490b-899b-e14bcd599eb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:28 crc kubenswrapper[4758]: I1203 17:14:28.892477 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:14:29 crc kubenswrapper[4758]: I1203 17:14:29.543338 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:29 crc kubenswrapper[4758]: I1203 17:14:29.543467 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:29 crc kubenswrapper[4758]: E1203 17:14:29.543907 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:14:29 crc kubenswrapper[4758]: E1203 17:14:29.544206 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs podName:ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c nodeName:}" failed. No retries permitted until 2025-12-03 17:14:45.544123857 +0000 UTC m=+1140.745500718 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs") pod "openstack-operator-controller-manager-84685d4b7d-4f4l6" (UID: "ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c") : secret "webhook-server-cert" not found Dec 03 17:14:29 crc kubenswrapper[4758]: I1203 17:14:29.550042 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-metrics-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:31 crc kubenswrapper[4758]: E1203 17:14:31.226380 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 03 17:14:31 crc kubenswrapper[4758]: E1203 17:14:31.227346 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xxj28,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-5mgxc_openstack-operators(c624801b-66ee-4a15-866a-a65e7c69b2df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:33 crc kubenswrapper[4758]: E1203 17:14:33.230232 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 03 17:14:33 crc kubenswrapper[4758]: E1203 17:14:33.230858 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j8t6j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-q8ddt_openstack-operators(61019cc3-111f-4f59-8e43-64b5dfdd22c4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:39 crc kubenswrapper[4758]: E1203 17:14:39.529157 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 03 17:14:39 crc kubenswrapper[4758]: E1203 17:14:39.529848 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-24wcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-8ll9j_openstack-operators(99462bc7-4e24-40b1-8ce5-555c47c8305f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:41 crc kubenswrapper[4758]: I1203 17:14:41.394771 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:14:41 crc kubenswrapper[4758]: I1203 17:14:41.395223 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:14:42 crc kubenswrapper[4758]: E1203 17:14:42.404817 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 03 17:14:42 crc kubenswrapper[4758]: E1203 17:14:42.405226 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lz6tw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-rp8pj_openstack-operators(7806e6cc-eee5-4735-bae7-10ca5e4144d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:43 crc kubenswrapper[4758]: E1203 17:14:43.629410 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 03 17:14:43 crc kubenswrapper[4758]: E1203 17:14:43.630064 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-69hc9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-54w5p_openstack-operators(56b5e5d0-a007-4451-bfbb-87e6f7201367): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:45 crc kubenswrapper[4758]: I1203 17:14:45.611813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:45 crc kubenswrapper[4758]: I1203 17:14:45.632450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c-webhook-certs\") pod \"openstack-operator-controller-manager-84685d4b7d-4f4l6\" (UID: \"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c\") " pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:45 crc kubenswrapper[4758]: E1203 17:14:45.680353 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 03 17:14:45 crc kubenswrapper[4758]: E1203 17:14:45.680660 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v22pt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-fnlpm_openstack-operators(de31e4e2-fe86-4543-82c5-da042e77b6ce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:45 crc kubenswrapper[4758]: I1203 17:14:45.710007 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-z7htx" Dec 03 17:14:45 crc kubenswrapper[4758]: I1203 17:14:45.717622 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:14:46 crc kubenswrapper[4758]: E1203 17:14:46.978818 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 03 17:14:46 crc kubenswrapper[4758]: E1203 17:14:46.979230 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k68r6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-72tmg_openstack-operators(84cea5a5-c77f-4818-9c41-8d91a838f7fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:49 crc kubenswrapper[4758]: E1203 17:14:49.500567 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 17:14:49 crc kubenswrapper[4758]: E1203 17:14:49.500943 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jzhpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-ndmwx_openstack-operators(74b430d9-3316-4157-ab4f-264f7797c7e9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:50 crc kubenswrapper[4758]: E1203 17:14:50.964539 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 03 17:14:50 crc kubenswrapper[4758]: E1203 17:14:50.965240 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxpdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-qcpt6_openstack-operators(f6893abc-6551-4bc5-bcc5-18dd74da025b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:51 crc kubenswrapper[4758]: E1203 17:14:51.548095 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 03 17:14:51 crc kubenswrapper[4758]: E1203 17:14:51.548348 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tjgrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-5djcl_openstack-operators(144cd5a6-c5c5-456d-b28d-27c3a3fb8898): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:52 crc kubenswrapper[4758]: E1203 17:14:52.627648 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 03 17:14:52 crc kubenswrapper[4758]: E1203 17:14:52.627929 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hhnnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-4972h_openstack-operators(5477637e-ea68-454e-8cf6-51b491464daf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:55 crc kubenswrapper[4758]: E1203 17:14:55.076016 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 03 17:14:55 crc kubenswrapper[4758]: E1203 17:14:55.076809 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-62tvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4pfjv_openstack-operators(aae66ae2-507f-43f8-bac5-c842427bb780): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:55 crc kubenswrapper[4758]: E1203 17:14:55.506597 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 03 17:14:55 crc kubenswrapper[4758]: E1203 17:14:55.506850 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-csf54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-6llwb_openstack-operators(68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:56 crc kubenswrapper[4758]: E1203 17:14:56.770766 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 17:14:56 crc kubenswrapper[4758]: E1203 17:14:56.771358 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j8t6j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-q8ddt_openstack-operators(61019cc3-111f-4f59-8e43-64b5dfdd22c4): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 17:14:56 crc kubenswrapper[4758]: E1203 17:14:56.772975 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" podUID="61019cc3-111f-4f59-8e43-64b5dfdd22c4" Dec 03 17:14:58 crc kubenswrapper[4758]: E1203 17:14:58.810448 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 17:14:58 crc kubenswrapper[4758]: E1203 17:14:58.810798 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g9wrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-c5wfm_openstack-operators(29c8457a-abba-43c0-b2ac-12c78bdeef4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:14:59 crc kubenswrapper[4758]: E1203 17:14:59.758442 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 17:14:59 crc kubenswrapper[4758]: E1203 17:14:59.761024 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-24wcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-8ll9j_openstack-operators(99462bc7-4e24-40b1-8ce5-555c47c8305f): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 17:14:59 crc kubenswrapper[4758]: E1203 17:14:59.762248 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" podUID="99462bc7-4e24-40b1-8ce5-555c47c8305f" Dec 03 17:14:59 crc kubenswrapper[4758]: E1203 17:14:59.906151 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 17:14:59 crc kubenswrapper[4758]: E1203 17:14:59.906429 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jt9b5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-r5r8j_openstack-operators(4ecb6824-b536-4432-8e37-92dc55964815): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.163308 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l"] Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.164875 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.168149 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.168441 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.173400 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l"] Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.190323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sddc8\" (UniqueName: \"kubernetes.io/projected/f93e2eca-7a66-4f9a-9151-7767b96dcc47-kube-api-access-sddc8\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.190986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f93e2eca-7a66-4f9a-9151-7767b96dcc47-config-volume\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.191035 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f93e2eca-7a66-4f9a-9151-7767b96dcc47-secret-volume\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.292994 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f93e2eca-7a66-4f9a-9151-7767b96dcc47-config-volume\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.293084 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f93e2eca-7a66-4f9a-9151-7767b96dcc47-secret-volume\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.293198 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sddc8\" (UniqueName: \"kubernetes.io/projected/f93e2eca-7a66-4f9a-9151-7767b96dcc47-kube-api-access-sddc8\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.294939 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f93e2eca-7a66-4f9a-9151-7767b96dcc47-config-volume\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.302512 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f93e2eca-7a66-4f9a-9151-7767b96dcc47-secret-volume\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.313763 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sddc8\" (UniqueName: \"kubernetes.io/projected/f93e2eca-7a66-4f9a-9151-7767b96dcc47-kube-api-access-sddc8\") pod \"collect-profiles-29413035-r4r7l\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: I1203 17:15:00.487841 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:00 crc kubenswrapper[4758]: E1203 17:15:00.887114 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 17:15:00 crc kubenswrapper[4758]: E1203 17:15:00.887401 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jbwz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-4vmjv_openstack-operators(5a38f83d-d017-49fc-98bf-e08ae999428c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:15:00 crc kubenswrapper[4758]: E1203 17:15:00.888562 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" podUID="5a38f83d-d017-49fc-98bf-e08ae999428c" Dec 03 17:15:01 crc kubenswrapper[4758]: I1203 17:15:01.297641 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j"] Dec 03 17:15:01 crc kubenswrapper[4758]: W1203 17:15:01.337550 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff85f702_0ab2_490b_899b_e14bcd599eb4.slice/crio-10a16c23dddab7ee0fceee8a3c8b187f2049dff317986bc4409fffc3bbc129f6 WatchSource:0}: Error finding container 10a16c23dddab7ee0fceee8a3c8b187f2049dff317986bc4409fffc3bbc129f6: Status 404 returned error can't find the container with id 10a16c23dddab7ee0fceee8a3c8b187f2049dff317986bc4409fffc3bbc129f6 Dec 03 17:15:01 crc kubenswrapper[4758]: I1203 17:15:01.445619 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" event={"ID":"ff85f702-0ab2-490b-899b-e14bcd599eb4","Type":"ContainerStarted","Data":"10a16c23dddab7ee0fceee8a3c8b187f2049dff317986bc4409fffc3bbc129f6"} Dec 03 17:15:01 crc kubenswrapper[4758]: I1203 17:15:01.459691 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-7m28f"] Dec 03 17:15:01 crc kubenswrapper[4758]: I1203 17:15:01.821965 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6"] Dec 03 17:15:03 crc kubenswrapper[4758]: E1203 17:15:03.208144 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 17:15:03 crc kubenswrapper[4758]: E1203 17:15:03.208824 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ssbtr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-2j99q_openstack-operators(eefa346f-ce5f-4445-8565-f6ad8250f049): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:15:03 crc kubenswrapper[4758]: W1203 17:15:03.209480 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee20d8c6_5b9b_4918_ac2c_4f30ca94c37c.slice/crio-5bd9a310433c26cf38b582661d9d44e81e967297da371e85f31ce906a89a5dfa WatchSource:0}: Error finding container 5bd9a310433c26cf38b582661d9d44e81e967297da371e85f31ce906a89a5dfa: Status 404 returned error can't find the container with id 5bd9a310433c26cf38b582661d9d44e81e967297da371e85f31ce906a89a5dfa Dec 03 17:15:03 crc kubenswrapper[4758]: E1203 17:15:03.213103 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" podUID="eefa346f-ce5f-4445-8565-f6ad8250f049" Dec 03 17:15:03 crc kubenswrapper[4758]: I1203 17:15:03.480068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" event={"ID":"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c","Type":"ContainerStarted","Data":"5bd9a310433c26cf38b582661d9d44e81e967297da371e85f31ce906a89a5dfa"} Dec 03 17:15:03 crc kubenswrapper[4758]: I1203 17:15:03.482230 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" event={"ID":"1911d89e-c35c-4104-b0c1-265f770a731d","Type":"ContainerStarted","Data":"19a0a91c3015c475c62024d09d09ae2c9702bb196638abeafb9320ddc98149af"} Dec 03 17:15:03 crc kubenswrapper[4758]: E1203 17:15:03.592809 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" podUID="61019cc3-111f-4f59-8e43-64b5dfdd22c4" Dec 03 17:15:03 crc kubenswrapper[4758]: E1203 17:15:03.592999 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" podUID="99462bc7-4e24-40b1-8ce5-555c47c8305f" Dec 03 17:15:03 crc kubenswrapper[4758]: I1203 17:15:03.723125 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l"] Dec 03 17:15:03 crc kubenswrapper[4758]: W1203 17:15:03.749038 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf93e2eca_7a66_4f9a_9151_7767b96dcc47.slice/crio-8f3dac109c8719441a59d2bcab8f46f5afe9c2d04dd7e503d3fbc76ab540b2a3 WatchSource:0}: Error finding container 8f3dac109c8719441a59d2bcab8f46f5afe9c2d04dd7e503d3fbc76ab540b2a3: Status 404 returned error can't find the container with id 8f3dac109c8719441a59d2bcab8f46f5afe9c2d04dd7e503d3fbc76ab540b2a3 Dec 03 17:15:03 crc kubenswrapper[4758]: E1203 17:15:03.842999 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" podUID="84cea5a5-c77f-4818-9c41-8d91a838f7fa" Dec 03 17:15:04 crc kubenswrapper[4758]: E1203 17:15:04.470630 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" podUID="68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.494608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" event={"ID":"61d9456a-70b6-4a3c-80fe-1ad9a65d0f44","Type":"ContainerStarted","Data":"48543a9209eede414138172fce8e87b5d488f692e14853cdaaa9b13cc5d393e0"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.494741 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" event={"ID":"61d9456a-70b6-4a3c-80fe-1ad9a65d0f44","Type":"ContainerStarted","Data":"452afc6152590b68f12ed8b999772b833f6b158887ffb7f69c6ef46967662f70"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.496009 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.497889 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" event={"ID":"61019cc3-111f-4f59-8e43-64b5dfdd22c4","Type":"ContainerStarted","Data":"1f85d0b0aba679ab225dff1aef84da36337cfcdb2dabea9f0fda7f7377eb5293"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.498523 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.501529 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" event={"ID":"ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c","Type":"ContainerStarted","Data":"9006b48555f6b420ffd6cf39dd589be3e6a3d9f2fad70499a4e6447d0e652342"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.501717 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.511442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" event={"ID":"f93e2eca-7a66-4f9a-9151-7767b96dcc47","Type":"ContainerStarted","Data":"8e5a54860065889dbf42cc1aef9277736db5b1da160cd695ee541ff59945aadc"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.511504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" event={"ID":"f93e2eca-7a66-4f9a-9151-7767b96dcc47","Type":"ContainerStarted","Data":"8f3dac109c8719441a59d2bcab8f46f5afe9c2d04dd7e503d3fbc76ab540b2a3"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.521458 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" podStartSLOduration=11.054894207 podStartE2EDuration="52.521420853s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.243572892 +0000 UTC m=+1110.444949763" lastFinishedPulling="2025-12-03 17:14:56.710099548 +0000 UTC m=+1151.911476409" observedRunningTime="2025-12-03 17:15:04.521222427 +0000 UTC m=+1159.722599288" watchObservedRunningTime="2025-12-03 17:15:04.521420853 +0000 UTC m=+1159.722797714" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.522293 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" event={"ID":"ce961712-a4d4-4a93-96e6-0cf4fdc58eea","Type":"ContainerStarted","Data":"d007b183208565c35b69ed650b0881b11ba32774806a71494e0a9ab74f16b5ba"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.528575 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" event={"ID":"68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1","Type":"ContainerStarted","Data":"81ee93e8d5a9d92f2401f68f0044a7dad3b9d17a6ebd7ac0e066a1209dca8b50"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.544455 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" event={"ID":"36504cff-c657-485e-9db7-788382c4a9e1","Type":"ContainerStarted","Data":"6d9d0084df8de086c6dccd535587cd5b4499cc006365a29910a59a5fe10680c7"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.544529 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" event={"ID":"36504cff-c657-485e-9db7-788382c4a9e1","Type":"ContainerStarted","Data":"9dd1c2ec671e94f6bacbf65bb459ff76dcc8fa96f65b4eb8c1a9a0908b48b967"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.544663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.551625 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" event={"ID":"84cea5a5-c77f-4818-9c41-8d91a838f7fa","Type":"ContainerStarted","Data":"bac9071c5a26769ec902c84bd222215256d3f756af8c8c31bf4ba3ddef11558a"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.558178 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" event={"ID":"99462bc7-4e24-40b1-8ce5-555c47c8305f","Type":"ContainerStarted","Data":"351882ee735564421b5f797bd48a7d992d11eee5a4fac0d67c0eaacc276a242c"} Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.559312 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.597933 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" podStartSLOduration=51.59789891 podStartE2EDuration="51.59789891s" podCreationTimestamp="2025-12-03 17:14:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:04.592246126 +0000 UTC m=+1159.793622997" watchObservedRunningTime="2025-12-03 17:15:04.59789891 +0000 UTC m=+1159.799275771" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.621919 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" podStartSLOduration=4.621894125 podStartE2EDuration="4.621894125s" podCreationTimestamp="2025-12-03 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:04.614560775 +0000 UTC m=+1159.815937646" watchObservedRunningTime="2025-12-03 17:15:04.621894125 +0000 UTC m=+1159.823270986" Dec 03 17:15:04 crc kubenswrapper[4758]: I1203 17:15:04.713224 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" podStartSLOduration=9.627067826 podStartE2EDuration="52.713199787s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:13.622344653 +0000 UTC m=+1108.823721514" lastFinishedPulling="2025-12-03 17:14:56.708476614 +0000 UTC m=+1151.909853475" observedRunningTime="2025-12-03 17:15:04.707663246 +0000 UTC m=+1159.909040107" watchObservedRunningTime="2025-12-03 17:15:04.713199787 +0000 UTC m=+1159.914576648" Dec 03 17:15:05 crc kubenswrapper[4758]: E1203 17:15:05.880005 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" podUID="68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1" Dec 03 17:15:06 crc kubenswrapper[4758]: I1203 17:15:06.581206 4758 generic.go:334] "Generic (PLEG): container finished" podID="f93e2eca-7a66-4f9a-9151-7767b96dcc47" containerID="8e5a54860065889dbf42cc1aef9277736db5b1da160cd695ee541ff59945aadc" exitCode=0 Dec 03 17:15:06 crc kubenswrapper[4758]: I1203 17:15:06.581256 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" event={"ID":"f93e2eca-7a66-4f9a-9151-7767b96dcc47","Type":"ContainerDied","Data":"8e5a54860065889dbf42cc1aef9277736db5b1da160cd695ee541ff59945aadc"} Dec 03 17:15:08 crc kubenswrapper[4758]: I1203 17:15:08.960485 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.140704 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f93e2eca-7a66-4f9a-9151-7767b96dcc47-secret-volume\") pod \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.141301 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sddc8\" (UniqueName: \"kubernetes.io/projected/f93e2eca-7a66-4f9a-9151-7767b96dcc47-kube-api-access-sddc8\") pod \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.141509 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f93e2eca-7a66-4f9a-9151-7767b96dcc47-config-volume\") pod \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\" (UID: \"f93e2eca-7a66-4f9a-9151-7767b96dcc47\") " Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.143102 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f93e2eca-7a66-4f9a-9151-7767b96dcc47-config-volume" (OuterVolumeSpecName: "config-volume") pod "f93e2eca-7a66-4f9a-9151-7767b96dcc47" (UID: "f93e2eca-7a66-4f9a-9151-7767b96dcc47"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.150308 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f93e2eca-7a66-4f9a-9151-7767b96dcc47-kube-api-access-sddc8" (OuterVolumeSpecName: "kube-api-access-sddc8") pod "f93e2eca-7a66-4f9a-9151-7767b96dcc47" (UID: "f93e2eca-7a66-4f9a-9151-7767b96dcc47"). InnerVolumeSpecName "kube-api-access-sddc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.151258 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93e2eca-7a66-4f9a-9151-7767b96dcc47-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f93e2eca-7a66-4f9a-9151-7767b96dcc47" (UID: "f93e2eca-7a66-4f9a-9151-7767b96dcc47"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.178434 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f93e2eca-7a66-4f9a-9151-7767b96dcc47-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.178661 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sddc8\" (UniqueName: \"kubernetes.io/projected/f93e2eca-7a66-4f9a-9151-7767b96dcc47-kube-api-access-sddc8\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.179049 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f93e2eca-7a66-4f9a-9151-7767b96dcc47-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.649184 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" event={"ID":"f93e2eca-7a66-4f9a-9151-7767b96dcc47","Type":"ContainerDied","Data":"8f3dac109c8719441a59d2bcab8f46f5afe9c2d04dd7e503d3fbc76ab540b2a3"} Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.649237 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f3dac109c8719441a59d2bcab8f46f5afe9c2d04dd7e503d3fbc76ab540b2a3" Dec 03 17:15:09 crc kubenswrapper[4758]: I1203 17:15:09.649325 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l" Dec 03 17:15:10 crc kubenswrapper[4758]: E1203 17:15:10.771513 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" podUID="29c8457a-abba-43c0-b2ac-12c78bdeef4c" Dec 03 17:15:10 crc kubenswrapper[4758]: E1203 17:15:10.911958 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" podUID="de31e4e2-fe86-4543-82c5-da042e77b6ce" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.070020 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" podUID="aae66ae2-507f-43f8-bac5-c842427bb780" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.182144 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" podUID="4ecb6824-b536-4432-8e37-92dc55964815" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.388577 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" podUID="c624801b-66ee-4a15-866a-a65e7c69b2df" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.394725 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.394791 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.526358 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" podUID="f6893abc-6551-4bc5-bcc5-18dd74da025b" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.526446 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" podUID="74b430d9-3316-4157-ab4f-264f7797c7e9" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.526536 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" podUID="5477637e-ea68-454e-8cf6-51b491464daf" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.526609 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" podUID="144cd5a6-c5c5-456d-b28d-27c3a3fb8898" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.760922 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" event={"ID":"de31e4e2-fe86-4543-82c5-da042e77b6ce","Type":"ContainerStarted","Data":"453210bf97bd10e9dbb8facec45afefb5f41073f9b99afb1f6ed2972bab9527a"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.767812 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" event={"ID":"61019cc3-111f-4f59-8e43-64b5dfdd22c4","Type":"ContainerStarted","Data":"a3c685ecef72f35e5312c667e7d98988c4663f2f485097a0981f9494f4fb9184"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.779101 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" event={"ID":"aae66ae2-507f-43f8-bac5-c842427bb780","Type":"ContainerStarted","Data":"6c29c021418c09ee41e929cf4676a7db921384b772feff62cdf2406d425c351a"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.780599 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" Dec 03 17:15:11 crc kubenswrapper[4758]: E1203 17:15:11.781758 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" podUID="aae66ae2-507f-43f8-bac5-c842427bb780" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.788060 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" event={"ID":"ff85f702-0ab2-490b-899b-e14bcd599eb4","Type":"ContainerStarted","Data":"6a4259a13964ced21a7cdb38eab103d94360e0c4ea652d15d5c44b800092020d"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.797053 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" event={"ID":"99462bc7-4e24-40b1-8ce5-555c47c8305f","Type":"ContainerStarted","Data":"4dc8cf2a21d4b8390462dd0b3fe534bb908b4af36bb28fb05f47a6eec9356534"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.804361 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.815121 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" event={"ID":"c624801b-66ee-4a15-866a-a65e7c69b2df","Type":"ContainerStarted","Data":"131407a89ba00958f47118b7ae4bec71bec0b382e19e4c68fb36c06f08c145b7"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.828531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" event={"ID":"84cea5a5-c77f-4818-9c41-8d91a838f7fa","Type":"ContainerStarted","Data":"8a20fa65af2d95dd224663b6b2b2f7bf91d77f383bbdf6b12a15384fc69e097b"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.829883 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.841983 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" event={"ID":"eefa346f-ce5f-4445-8565-f6ad8250f049","Type":"ContainerStarted","Data":"8da1808ff69afcff9ee9c5517e30f8456020877c1e570f61d9b69a98c446654c"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.849585 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-q8ddt" podStartSLOduration=11.265091846 podStartE2EDuration="59.849564113s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:14.645294903 +0000 UTC m=+1109.846671764" lastFinishedPulling="2025-12-03 17:15:03.22976717 +0000 UTC m=+1158.431144031" observedRunningTime="2025-12-03 17:15:11.826943298 +0000 UTC m=+1167.028320159" watchObservedRunningTime="2025-12-03 17:15:11.849564113 +0000 UTC m=+1167.050940974" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.851730 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" event={"ID":"f6893abc-6551-4bc5-bcc5-18dd74da025b","Type":"ContainerStarted","Data":"3c93e75c15fb3c32aa59be339d43d361791002ed64f89831439122ce37e68d76"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.877143 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" event={"ID":"29c8457a-abba-43c0-b2ac-12c78bdeef4c","Type":"ContainerStarted","Data":"9b03569fa8e01c92f792cfaa59c57bd199fc5a995678793132a2a56379d1ab02"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.890483 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" event={"ID":"4ecb6824-b536-4432-8e37-92dc55964815","Type":"ContainerStarted","Data":"62ec8ed3c3e0e4ee134fd3efc16566c1622069474960cf100d2f34d8808c13ed"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.915293 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" event={"ID":"1911d89e-c35c-4104-b0c1-265f770a731d","Type":"ContainerStarted","Data":"bb5bf529c275dfd0084d7e822ee2b709edb81a145f983c079110a60d848f96b1"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.929051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" event={"ID":"74b430d9-3316-4157-ab4f-264f7797c7e9","Type":"ContainerStarted","Data":"df25e7a9d56ac822950b16cc42090e4d3326bd60d94d018c940b4cb83a70be06"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.930588 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" podStartSLOduration=3.315162811 podStartE2EDuration="59.930553796s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:13.782475693 +0000 UTC m=+1108.983852564" lastFinishedPulling="2025-12-03 17:15:10.397866688 +0000 UTC m=+1165.599243549" observedRunningTime="2025-12-03 17:15:11.919523086 +0000 UTC m=+1167.120899967" watchObservedRunningTime="2025-12-03 17:15:11.930553796 +0000 UTC m=+1167.131930677" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.943395 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" event={"ID":"5477637e-ea68-454e-8cf6-51b491464daf","Type":"ContainerStarted","Data":"7e72a0d47e1fc23c21b1bc574080e308d19a2a61c6a38e9c8f6fe33bc92c2629"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.961938 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" event={"ID":"ce961712-a4d4-4a93-96e6-0cf4fdc58eea","Type":"ContainerStarted","Data":"9e9b3b08a2b475d266409786521dc2ee7b902a95cdf5cff8adfba794d7432cd3"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.962694 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8ll9j" podStartSLOduration=11.906138511 podStartE2EDuration="59.962660299s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.190225605 +0000 UTC m=+1110.391602466" lastFinishedPulling="2025-12-03 17:15:03.246747393 +0000 UTC m=+1158.448124254" observedRunningTime="2025-12-03 17:15:11.953563221 +0000 UTC m=+1167.154940082" watchObservedRunningTime="2025-12-03 17:15:11.962660299 +0000 UTC m=+1167.164037160" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.963399 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.969661 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" event={"ID":"144cd5a6-c5c5-456d-b28d-27c3a3fb8898","Type":"ContainerStarted","Data":"bc0360ae541a744357a15c6ee634f769d72aca7ea4391175c53a0669212186ed"} Dec 03 17:15:11 crc kubenswrapper[4758]: I1203 17:15:11.994427 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" Dec 03 17:15:12 crc kubenswrapper[4758]: I1203 17:15:12.058901 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-p8mlv" podStartSLOduration=4.827475474 podStartE2EDuration="1m0.058877225s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.205160183 +0000 UTC m=+1110.406537044" lastFinishedPulling="2025-12-03 17:15:10.436561934 +0000 UTC m=+1165.637938795" observedRunningTime="2025-12-03 17:15:12.05061062 +0000 UTC m=+1167.251987481" watchObservedRunningTime="2025-12-03 17:15:12.058877225 +0000 UTC m=+1167.260254086" Dec 03 17:15:12 crc kubenswrapper[4758]: E1203 17:15:12.071090 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" podUID="56b5e5d0-a007-4451-bfbb-87e6f7201367" Dec 03 17:15:12 crc kubenswrapper[4758]: E1203 17:15:12.108279 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" podUID="7806e6cc-eee5-4735-bae7-10ca5e4144d6" Dec 03 17:15:12 crc kubenswrapper[4758]: E1203 17:15:12.124908 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" podUID="5a38f83d-d017-49fc-98bf-e08ae999428c" Dec 03 17:15:12 crc kubenswrapper[4758]: I1203 17:15:12.709018 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4sfr" Dec 03 17:15:12 crc kubenswrapper[4758]: I1203 17:15:12.996194 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" event={"ID":"ff85f702-0ab2-490b-899b-e14bcd599eb4","Type":"ContainerStarted","Data":"9e2c160e85e3e28422a9fb3b40cbbf82f420e044e2b98aacd3662ef9aa69d99a"} Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:12.997171 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:12.999854 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" event={"ID":"56b5e5d0-a007-4451-bfbb-87e6f7201367","Type":"ContainerStarted","Data":"bd3d98c24e094f11ed7b103ab04f7bb7d6aad9455e849252401002bb779c3146"} Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.006535 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" event={"ID":"de31e4e2-fe86-4543-82c5-da042e77b6ce","Type":"ContainerStarted","Data":"96d83001eb361b6656e3b406767c23af66a352c020bccb744b5edb4f6efdf5e4"} Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.014938 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" event={"ID":"7806e6cc-eee5-4735-bae7-10ca5e4144d6","Type":"ContainerStarted","Data":"ea91f4188a2031326ae6425816af3e7962d0c24a9211722f622459ace87c9387"} Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.021448 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" event={"ID":"1911d89e-c35c-4104-b0c1-265f770a731d","Type":"ContainerStarted","Data":"1b2ae3c60484294f602f27634ddebc30b72e713d283d0234b03e69da300f7b66"} Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.022005 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.035805 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" event={"ID":"eefa346f-ce5f-4445-8565-f6ad8250f049","Type":"ContainerStarted","Data":"5780385151fb4aa9da9569194076f2585e21b37874b4d22c9afe6b3d982726b3"} Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.035843 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.045880 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" podStartSLOduration=52.145456005 podStartE2EDuration="1m1.045850072s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:15:01.360443441 +0000 UTC m=+1156.561820302" lastFinishedPulling="2025-12-03 17:15:10.260837508 +0000 UTC m=+1165.462214369" observedRunningTime="2025-12-03 17:15:13.032027746 +0000 UTC m=+1168.233404627" watchObservedRunningTime="2025-12-03 17:15:13.045850072 +0000 UTC m=+1168.247226933" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.078512 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" podStartSLOduration=53.909360163 podStartE2EDuration="1m1.078496159s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:15:03.22095319 +0000 UTC m=+1158.422330051" lastFinishedPulling="2025-12-03 17:15:10.390089186 +0000 UTC m=+1165.591466047" observedRunningTime="2025-12-03 17:15:13.065852915 +0000 UTC m=+1168.267229766" watchObservedRunningTime="2025-12-03 17:15:13.078496159 +0000 UTC m=+1168.279873020" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.192866 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" podStartSLOduration=3.560908259 podStartE2EDuration="1m1.192843239s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:14.645704593 +0000 UTC m=+1109.847081454" lastFinishedPulling="2025-12-03 17:15:12.277639583 +0000 UTC m=+1167.479016434" observedRunningTime="2025-12-03 17:15:13.130846202 +0000 UTC m=+1168.332223083" watchObservedRunningTime="2025-12-03 17:15:13.192843239 +0000 UTC m=+1168.394220100" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.626646 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-5x7rp" Dec 03 17:15:13 crc kubenswrapper[4758]: I1203 17:15:13.661149 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" podStartSLOduration=6.469465995 podStartE2EDuration="1m1.661125731s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.243991623 +0000 UTC m=+1110.445368474" lastFinishedPulling="2025-12-03 17:15:10.435651349 +0000 UTC m=+1165.637028210" observedRunningTime="2025-12-03 17:15:13.244268676 +0000 UTC m=+1168.445645537" watchObservedRunningTime="2025-12-03 17:15:13.661125731 +0000 UTC m=+1168.862502582" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.054627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" event={"ID":"4ecb6824-b536-4432-8e37-92dc55964815","Type":"ContainerStarted","Data":"43bafa07dbba85cbfde6a2696a4ea935c969a63848decb3cb2afa54fe1b0a8e9"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.054803 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.062766 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" event={"ID":"c624801b-66ee-4a15-866a-a65e7c69b2df","Type":"ContainerStarted","Data":"c68d58e86872b208a54135a9bba9ffaa1e625d31a36ac8832eb726d1dbf771ac"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.063821 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.065942 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" event={"ID":"74b430d9-3316-4157-ab4f-264f7797c7e9","Type":"ContainerStarted","Data":"c71984d46fd7d559e3244f3647930afe67c550d19dd20de1d284982d9fd9ba0d"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.066545 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.068287 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" event={"ID":"5477637e-ea68-454e-8cf6-51b491464daf","Type":"ContainerStarted","Data":"690745d8ff88feb3d82f1ecf028acff5c23864082867e915f917999f680798a4"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.068840 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.070731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" event={"ID":"f6893abc-6551-4bc5-bcc5-18dd74da025b","Type":"ContainerStarted","Data":"372d343e3090400e5e5c8d7e527d2543971340f75991aaaa64a6029f28c8d60e"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.070885 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.085241 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" event={"ID":"29c8457a-abba-43c0-b2ac-12c78bdeef4c","Type":"ContainerStarted","Data":"ae2543b1317c2501c27448d0061ec698e6d3fb0ec29bfe4acad536bacbd90369"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.086013 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.087998 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" event={"ID":"144cd5a6-c5c5-456d-b28d-27c3a3fb8898","Type":"ContainerStarted","Data":"4a4cefc077c667d7ea304d76ff4e63b55f8d48091408d9b967ff3337d89f7545"} Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.088813 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.089120 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.099025 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" podStartSLOduration=4.664855382 podStartE2EDuration="1m2.099006638s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.185776774 +0000 UTC m=+1110.387153635" lastFinishedPulling="2025-12-03 17:15:12.61992803 +0000 UTC m=+1167.821304891" observedRunningTime="2025-12-03 17:15:14.095241285 +0000 UTC m=+1169.296618146" watchObservedRunningTime="2025-12-03 17:15:14.099006638 +0000 UTC m=+1169.300383499" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.120267 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" podStartSLOduration=5.181383657 podStartE2EDuration="1m2.120251255s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.68318869 +0000 UTC m=+1110.884565551" lastFinishedPulling="2025-12-03 17:15:12.622056288 +0000 UTC m=+1167.823433149" observedRunningTime="2025-12-03 17:15:14.118353124 +0000 UTC m=+1169.319729985" watchObservedRunningTime="2025-12-03 17:15:14.120251255 +0000 UTC m=+1169.321628116" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.249394 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" podStartSLOduration=4.972772615 podStartE2EDuration="1m2.249357256s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.198752668 +0000 UTC m=+1110.400129539" lastFinishedPulling="2025-12-03 17:15:12.475337319 +0000 UTC m=+1167.676714180" observedRunningTime="2025-12-03 17:15:14.237051861 +0000 UTC m=+1169.438428732" watchObservedRunningTime="2025-12-03 17:15:14.249357256 +0000 UTC m=+1169.450734127" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.351037 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" podStartSLOduration=4.7629161 podStartE2EDuration="1m2.35099577s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.198952504 +0000 UTC m=+1110.400329365" lastFinishedPulling="2025-12-03 17:15:12.787032174 +0000 UTC m=+1167.988409035" observedRunningTime="2025-12-03 17:15:14.3477067 +0000 UTC m=+1169.549083561" watchObservedRunningTime="2025-12-03 17:15:14.35099577 +0000 UTC m=+1169.552372631" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.386901 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" podStartSLOduration=4.410055478 podStartE2EDuration="1m2.386869734s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:14.64593104 +0000 UTC m=+1109.847307901" lastFinishedPulling="2025-12-03 17:15:12.622745306 +0000 UTC m=+1167.824122157" observedRunningTime="2025-12-03 17:15:14.378594869 +0000 UTC m=+1169.579971750" watchObservedRunningTime="2025-12-03 17:15:14.386869734 +0000 UTC m=+1169.588246595" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.406024 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" podStartSLOduration=5.26776425 podStartE2EDuration="1m2.406002225s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.141726242 +0000 UTC m=+1110.343103103" lastFinishedPulling="2025-12-03 17:15:12.279964217 +0000 UTC m=+1167.481341078" observedRunningTime="2025-12-03 17:15:14.402251703 +0000 UTC m=+1169.603628564" watchObservedRunningTime="2025-12-03 17:15:14.406002225 +0000 UTC m=+1169.607379086" Dec 03 17:15:14 crc kubenswrapper[4758]: I1203 17:15:14.691113 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" podStartSLOduration=4.622305843 podStartE2EDuration="1m2.691077977s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:14.674373896 +0000 UTC m=+1109.875750757" lastFinishedPulling="2025-12-03 17:15:12.74314603 +0000 UTC m=+1167.944522891" observedRunningTime="2025-12-03 17:15:14.532779462 +0000 UTC m=+1169.734156393" watchObservedRunningTime="2025-12-03 17:15:14.691077977 +0000 UTC m=+1169.892454838" Dec 03 17:15:15 crc kubenswrapper[4758]: I1203 17:15:15.102443 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7m28f" Dec 03 17:15:15 crc kubenswrapper[4758]: I1203 17:15:15.724541 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-84685d4b7d-4f4l6" Dec 03 17:15:18 crc kubenswrapper[4758]: I1203 17:15:18.172351 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" event={"ID":"56b5e5d0-a007-4451-bfbb-87e6f7201367","Type":"ContainerStarted","Data":"8f889c4dd037a3c22de5a8f1c30b70cb8be571fb803d5038f00150dcb5d80f6e"} Dec 03 17:15:18 crc kubenswrapper[4758]: I1203 17:15:18.174257 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" event={"ID":"7806e6cc-eee5-4735-bae7-10ca5e4144d6","Type":"ContainerStarted","Data":"8198d0a78e09ba90a8c2380e01af066e202d3c17a38d7108083997d5479a9760"} Dec 03 17:15:18 crc kubenswrapper[4758]: I1203 17:15:18.902990 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j" Dec 03 17:15:19 crc kubenswrapper[4758]: I1203 17:15:19.183086 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:15:19 crc kubenswrapper[4758]: I1203 17:15:19.206494 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" podStartSLOduration=5.731090415 podStartE2EDuration="1m7.206472414s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.692411472 +0000 UTC m=+1110.893788333" lastFinishedPulling="2025-12-03 17:15:17.167793471 +0000 UTC m=+1172.369170332" observedRunningTime="2025-12-03 17:15:19.204233443 +0000 UTC m=+1174.405610324" watchObservedRunningTime="2025-12-03 17:15:19.206472414 +0000 UTC m=+1174.407849275" Dec 03 17:15:19 crc kubenswrapper[4758]: I1203 17:15:19.224836 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" podStartSLOduration=5.2964140109999995 podStartE2EDuration="1m7.224813033s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.241945067 +0000 UTC m=+1110.443321928" lastFinishedPulling="2025-12-03 17:15:17.170344089 +0000 UTC m=+1172.371720950" observedRunningTime="2025-12-03 17:15:19.222336245 +0000 UTC m=+1174.423713106" watchObservedRunningTime="2025-12-03 17:15:19.224813033 +0000 UTC m=+1174.426189894" Dec 03 17:15:22 crc kubenswrapper[4758]: I1203 17:15:22.521256 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fnlpm" Dec 03 17:15:22 crc kubenswrapper[4758]: I1203 17:15:22.592645 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-72tmg" Dec 03 17:15:22 crc kubenswrapper[4758]: I1203 17:15:22.917625 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5djcl" Dec 03 17:15:22 crc kubenswrapper[4758]: I1203 17:15:22.923839 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4972h" Dec 03 17:15:22 crc kubenswrapper[4758]: I1203 17:15:22.988698 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-r5r8j" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.095460 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qcpt6" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.129804 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-5mgxc" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.132710 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-ndmwx" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.187826 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c5wfm" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.244011 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-rp8pj" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.536384 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" event={"ID":"68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1","Type":"ContainerStarted","Data":"6c9b1a1abbfb76ee069acfd74049d80e2ef04163605b0f4982aca27eb6de11b0"} Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.536758 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.557659 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" podStartSLOduration=4.158863699 podStartE2EDuration="1m11.557632036s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.243962212 +0000 UTC m=+1110.445339073" lastFinishedPulling="2025-12-03 17:15:22.642730549 +0000 UTC m=+1177.844107410" observedRunningTime="2025-12-03 17:15:23.552507637 +0000 UTC m=+1178.753884508" watchObservedRunningTime="2025-12-03 17:15:23.557632036 +0000 UTC m=+1178.759008897" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.669123 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2j99q" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.951893 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:15:23 crc kubenswrapper[4758]: I1203 17:15:23.954641 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-54w5p" Dec 03 17:15:28 crc kubenswrapper[4758]: I1203 17:15:28.580200 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" event={"ID":"5a38f83d-d017-49fc-98bf-e08ae999428c","Type":"ContainerStarted","Data":"6993c96d480bfa0597e9982b826c2a14b62855f849ebbfa039656e66bc1c4b49"} Dec 03 17:15:28 crc kubenswrapper[4758]: I1203 17:15:28.582455 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" event={"ID":"aae66ae2-507f-43f8-bac5-c842427bb780","Type":"ContainerStarted","Data":"a72240d867102529b4add44860cce91c4d8638a737ce82cc5f955510aa15facf"} Dec 03 17:15:28 crc kubenswrapper[4758]: I1203 17:15:28.582824 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:15:28 crc kubenswrapper[4758]: I1203 17:15:28.602844 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4vmjv" podStartSLOduration=2.932685839 podStartE2EDuration="1m15.602825249s" podCreationTimestamp="2025-12-03 17:14:13 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.243984443 +0000 UTC m=+1110.445361304" lastFinishedPulling="2025-12-03 17:15:27.914123853 +0000 UTC m=+1183.115500714" observedRunningTime="2025-12-03 17:15:28.600158467 +0000 UTC m=+1183.801535348" watchObservedRunningTime="2025-12-03 17:15:28.602825249 +0000 UTC m=+1183.804202110" Dec 03 17:15:28 crc kubenswrapper[4758]: I1203 17:15:28.629504 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" podStartSLOduration=4.283286331 podStartE2EDuration="1m16.629481824s" podCreationTimestamp="2025-12-03 17:14:12 +0000 UTC" firstStartedPulling="2025-12-03 17:14:15.706296811 +0000 UTC m=+1110.907673672" lastFinishedPulling="2025-12-03 17:15:28.052492304 +0000 UTC m=+1183.253869165" observedRunningTime="2025-12-03 17:15:28.622932036 +0000 UTC m=+1183.824308897" watchObservedRunningTime="2025-12-03 17:15:28.629481824 +0000 UTC m=+1183.830858685" Dec 03 17:15:33 crc kubenswrapper[4758]: I1203 17:15:33.621159 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4pfjv" Dec 03 17:15:33 crc kubenswrapper[4758]: I1203 17:15:33.770019 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-6llwb" Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.395594 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.397060 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.397127 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.397734 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74575219fc052d90d1263cdeb0cdc0186ad14da295a8ec4c660422997cca91c3"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.397795 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://74575219fc052d90d1263cdeb0cdc0186ad14da295a8ec4c660422997cca91c3" gracePeriod=600 Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.708580 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="74575219fc052d90d1263cdeb0cdc0186ad14da295a8ec4c660422997cca91c3" exitCode=0 Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.708649 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"74575219fc052d90d1263cdeb0cdc0186ad14da295a8ec4c660422997cca91c3"} Dec 03 17:15:41 crc kubenswrapper[4758]: I1203 17:15:41.708730 4758 scope.go:117] "RemoveContainer" containerID="8050d5b5272390ff909cae77aca4e99c2b09448e05fbe4bc6797c541002e26b0" Dec 03 17:15:43 crc kubenswrapper[4758]: I1203 17:15:43.740554 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"edeb1f97ccdd747653b82f053447c5a4a38ed821ec3f1cb524dc9338c014ae57"} Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.589249 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5wcgm"] Dec 03 17:15:49 crc kubenswrapper[4758]: E1203 17:15:49.594804 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93e2eca-7a66-4f9a-9151-7767b96dcc47" containerName="collect-profiles" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.595193 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93e2eca-7a66-4f9a-9151-7767b96dcc47" containerName="collect-profiles" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.595426 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93e2eca-7a66-4f9a-9151-7767b96dcc47" containerName="collect-profiles" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.596470 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.601604 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.601917 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-dknkv" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.602134 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.614569 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tzg7\" (UniqueName: \"kubernetes.io/projected/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-kube-api-access-4tzg7\") pod \"dnsmasq-dns-675f4bcbfc-5wcgm\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.614719 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-config\") pod \"dnsmasq-dns-675f4bcbfc-5wcgm\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.620436 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.627754 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5wcgm"] Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.705135 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tj6n2"] Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.706645 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.710007 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.716218 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzrvn\" (UniqueName: \"kubernetes.io/projected/656169ac-4510-47b8-961c-ae8d2439af94-kube-api-access-vzrvn\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.716324 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.716360 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-config\") pod \"dnsmasq-dns-675f4bcbfc-5wcgm\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.716398 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-config\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.716526 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tzg7\" (UniqueName: \"kubernetes.io/projected/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-kube-api-access-4tzg7\") pod \"dnsmasq-dns-675f4bcbfc-5wcgm\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.717744 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-config\") pod \"dnsmasq-dns-675f4bcbfc-5wcgm\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.729039 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tj6n2"] Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.761872 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tzg7\" (UniqueName: \"kubernetes.io/projected/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-kube-api-access-4tzg7\") pod \"dnsmasq-dns-675f4bcbfc-5wcgm\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.818017 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzrvn\" (UniqueName: \"kubernetes.io/projected/656169ac-4510-47b8-961c-ae8d2439af94-kube-api-access-vzrvn\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.818070 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.819224 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.819292 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-config\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.819671 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-config\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.840423 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzrvn\" (UniqueName: \"kubernetes.io/projected/656169ac-4510-47b8-961c-ae8d2439af94-kube-api-access-vzrvn\") pod \"dnsmasq-dns-78dd6ddcc-tj6n2\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:49 crc kubenswrapper[4758]: I1203 17:15:49.917288 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:15:50 crc kubenswrapper[4758]: I1203 17:15:50.026058 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:15:50 crc kubenswrapper[4758]: I1203 17:15:50.350082 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5wcgm"] Dec 03 17:15:50 crc kubenswrapper[4758]: I1203 17:15:50.676182 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tj6n2"] Dec 03 17:15:50 crc kubenswrapper[4758]: W1203 17:15:50.682018 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod656169ac_4510_47b8_961c_ae8d2439af94.slice/crio-d51bb8ac07e6f17cb097dfb12dbea6c8c80d9d6df0772fa127aa85dd37e48445 WatchSource:0}: Error finding container d51bb8ac07e6f17cb097dfb12dbea6c8c80d9d6df0772fa127aa85dd37e48445: Status 404 returned error can't find the container with id d51bb8ac07e6f17cb097dfb12dbea6c8c80d9d6df0772fa127aa85dd37e48445 Dec 03 17:15:50 crc kubenswrapper[4758]: I1203 17:15:50.830350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" event={"ID":"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992","Type":"ContainerStarted","Data":"01e16815e497d58d1a41a615a72a7d0adc7710b037aa4b1a795f513bebeed9d5"} Dec 03 17:15:50 crc kubenswrapper[4758]: I1203 17:15:50.831995 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" event={"ID":"656169ac-4510-47b8-961c-ae8d2439af94","Type":"ContainerStarted","Data":"d51bb8ac07e6f17cb097dfb12dbea6c8c80d9d6df0772fa127aa85dd37e48445"} Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.194614 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5wcgm"] Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.336718 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p9z46"] Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.339131 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.429066 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.429186 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8mns\" (UniqueName: \"kubernetes.io/projected/cc544ec4-59b0-4acf-a808-db315c913f98-kube-api-access-m8mns\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.429276 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-config\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.481696 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p9z46"] Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.531555 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.531643 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8mns\" (UniqueName: \"kubernetes.io/projected/cc544ec4-59b0-4acf-a808-db315c913f98-kube-api-access-m8mns\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.531732 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-config\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.532912 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.532999 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-config\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.573379 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8mns\" (UniqueName: \"kubernetes.io/projected/cc544ec4-59b0-4acf-a808-db315c913f98-kube-api-access-m8mns\") pod \"dnsmasq-dns-666b6646f7-p9z46\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.691585 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.861937 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tj6n2"] Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.896275 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5v42"] Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.898066 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:52 crc kubenswrapper[4758]: I1203 17:15:52.916016 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5v42"] Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.075193 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt5km\" (UniqueName: \"kubernetes.io/projected/cdd798d4-66a3-4087-a7c8-bd1bbc101664-kube-api-access-bt5km\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.075271 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-config\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.075332 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.176781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt5km\" (UniqueName: \"kubernetes.io/projected/cdd798d4-66a3-4087-a7c8-bd1bbc101664-kube-api-access-bt5km\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.176865 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-config\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.176914 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.180013 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-config\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.180608 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.210226 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt5km\" (UniqueName: \"kubernetes.io/projected/cdd798d4-66a3-4087-a7c8-bd1bbc101664-kube-api-access-bt5km\") pod \"dnsmasq-dns-57d769cc4f-c5v42\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.271335 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.505205 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.510004 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.518351 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.518673 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5g26j" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.518936 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.530184 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.532042 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.532275 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.532884 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.539850 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.689550 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.689704 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/070406cc-6d65-4fd0-81d7-264adead6bbd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.689782 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.689855 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.689897 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gk9r\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-kube-api-access-6gk9r\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.689963 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.690028 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.690073 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.690107 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/070406cc-6d65-4fd0-81d7-264adead6bbd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.690145 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.690185 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.805340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.806079 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.806165 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.806208 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/070406cc-6d65-4fd0-81d7-264adead6bbd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.806288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.806385 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.807346 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.807540 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/070406cc-6d65-4fd0-81d7-264adead6bbd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.807722 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.807833 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.807866 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gk9r\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-kube-api-access-6gk9r\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.813026 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.815541 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.816432 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.817225 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.818145 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.824706 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.828298 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/070406cc-6d65-4fd0-81d7-264adead6bbd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.836975 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.839997 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/070406cc-6d65-4fd0-81d7-264adead6bbd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.842417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:53 crc kubenswrapper[4758]: I1203 17:15:53.852181 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gk9r\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-kube-api-access-6gk9r\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.002093 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " pod="openstack/rabbitmq-server-0" Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.257031 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.404749 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p9z46"] Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.565066 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5v42"] Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.803285 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.947310 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"070406cc-6d65-4fd0-81d7-264adead6bbd","Type":"ContainerStarted","Data":"4d97514319b93120ec43aea8688d1f35b4a402fc2d7314bc55922c9f682aa8e1"} Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.949657 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" event={"ID":"cc544ec4-59b0-4acf-a808-db315c913f98","Type":"ContainerStarted","Data":"47e6f83d9af8d95ec8e103fee1b1ea568c43b40f6d33aa6c03578e4f424b3e6f"} Dec 03 17:15:54 crc kubenswrapper[4758]: I1203 17:15:54.952485 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" event={"ID":"cdd798d4-66a3-4087-a7c8-bd1bbc101664","Type":"ContainerStarted","Data":"c0ecb0a2b67c7ef56dc337a6ec25907155ad8ffb9b55b2f5e4471a539232dd18"} Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.018374 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.021071 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.027242 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2pl76" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.027586 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.029471 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.029601 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.029753 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.029892 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.032099 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.038135 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078356 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078483 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078539 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078607 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078640 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078704 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078732 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8mjg\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-kube-api-access-g8mjg\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078790 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e518390e-b4c4-49bb-913f-ddfc331456ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.078940 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e518390e-b4c4-49bb-913f-ddfc331456ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.106523 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.108454 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.114656 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.116938 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-n9fv6" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.117271 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.117353 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.122965 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.166391 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245216 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245277 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245301 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245328 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245345 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245360 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245376 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8mjg\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-kube-api-access-g8mjg\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245394 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e518390e-b4c4-49bb-913f-ddfc331456ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245414 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245433 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e518390e-b4c4-49bb-913f-ddfc331456ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.245449 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.246554 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.251585 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.251852 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.252514 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.267762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.269101 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.278780 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e518390e-b4c4-49bb-913f-ddfc331456ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.296707 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8mjg\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-kube-api-access-g8mjg\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.322156 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.322375 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.330297 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e518390e-b4c4-49bb-913f-ddfc331456ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.350785 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kolla-config\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.350931 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.350957 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.350996 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.351018 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.351035 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-default\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.351058 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.351099 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccv6m\" (UniqueName: \"kubernetes.io/projected/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kube-api-access-ccv6m\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.369992 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.377279 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452529 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452552 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-default\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452612 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccv6m\" (UniqueName: \"kubernetes.io/projected/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kube-api-access-ccv6m\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452657 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kolla-config\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452726 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.452746 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.453811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.454647 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kolla-config\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.455044 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.455185 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.455465 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-default\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.465918 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.476405 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccv6m\" (UniqueName: \"kubernetes.io/projected/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kube-api-access-ccv6m\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.500668 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.524341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.758329 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 17:15:55 crc kubenswrapper[4758]: I1203 17:15:55.926431 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.018336 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e518390e-b4c4-49bb-913f-ddfc331456ab","Type":"ContainerStarted","Data":"184e063513f2f136edb7bc9d67f4581f9420630c7e2d1fd923601f13ea501aa5"} Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.457746 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.463298 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.470031 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-lcshp" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.470959 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.472098 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.473601 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.514853 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.621063 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.736447 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.736535 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.736562 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggl67\" (UniqueName: \"kubernetes.io/projected/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kube-api-access-ggl67\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.736598 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.737060 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.737222 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.737337 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.737401 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839278 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839449 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839471 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839524 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggl67\" (UniqueName: \"kubernetes.io/projected/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kube-api-access-ggl67\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.839581 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.841331 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.842008 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.848977 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.849668 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.850393 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.853410 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.862473 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.864075 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.877288 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.880434 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.880796 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-v75jr" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.896473 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.919307 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggl67\" (UniqueName: \"kubernetes.io/projected/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kube-api-access-ggl67\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.919418 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:56 crc kubenswrapper[4758]: I1203 17:15:56.933773 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.043580 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9c80de0e-2bee-4e3f-8edf-f4a235b01d76","Type":"ContainerStarted","Data":"9ba0dbb96f7bfc52b8dc940d806620d01eb3d778c6e4b13f0e71b9a472076e1e"} Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.063998 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.064086 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.064170 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-kolla-config\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.064193 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-config-data\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.064249 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w2dg\" (UniqueName: \"kubernetes.io/projected/0754ea69-5007-4e51-bd22-0efd13954199-kube-api-access-9w2dg\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.100431 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.169477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-kolla-config\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.169546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-config-data\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.169605 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w2dg\" (UniqueName: \"kubernetes.io/projected/0754ea69-5007-4e51-bd22-0efd13954199-kube-api-access-9w2dg\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.169711 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.169780 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.175824 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-kolla-config\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.178395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-config-data\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.192950 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.200066 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.202442 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w2dg\" (UniqueName: \"kubernetes.io/projected/0754ea69-5007-4e51-bd22-0efd13954199-kube-api-access-9w2dg\") pod \"memcached-0\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.299452 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 17:15:57 crc kubenswrapper[4758]: I1203 17:15:57.725093 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:15:58 crc kubenswrapper[4758]: I1203 17:15:58.074019 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"80b1161a-3f70-4956-b13a-1ef693aa2e7c","Type":"ContainerStarted","Data":"d6b24e0378e579d80ec113e94b696b20975fa5f080e73db1f5b7c7a8a17eee50"} Dec 03 17:15:58 crc kubenswrapper[4758]: I1203 17:15:58.151384 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.284737 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.297199 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0754ea69-5007-4e51-bd22-0efd13954199","Type":"ContainerStarted","Data":"fc7a474a0f49cd7b9fbb633251ad467314a7636da93a401cf9ae22068a10a4e3"} Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.297386 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.318488 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nx7pp" Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.357281 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.463362 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8njx\" (UniqueName: \"kubernetes.io/projected/d4bbd2d2-37a4-46d1-b487-1ece965b4a11-kube-api-access-f8njx\") pod \"kube-state-metrics-0\" (UID: \"d4bbd2d2-37a4-46d1-b487-1ece965b4a11\") " pod="openstack/kube-state-metrics-0" Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.564663 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8njx\" (UniqueName: \"kubernetes.io/projected/d4bbd2d2-37a4-46d1-b487-1ece965b4a11-kube-api-access-f8njx\") pod \"kube-state-metrics-0\" (UID: \"d4bbd2d2-37a4-46d1-b487-1ece965b4a11\") " pod="openstack/kube-state-metrics-0" Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.607644 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8njx\" (UniqueName: \"kubernetes.io/projected/d4bbd2d2-37a4-46d1-b487-1ece965b4a11-kube-api-access-f8njx\") pod \"kube-state-metrics-0\" (UID: \"d4bbd2d2-37a4-46d1-b487-1ece965b4a11\") " pod="openstack/kube-state-metrics-0" Dec 03 17:15:59 crc kubenswrapper[4758]: I1203 17:15:59.660137 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:16:00 crc kubenswrapper[4758]: I1203 17:16:00.950623 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:16:01 crc kubenswrapper[4758]: W1203 17:16:01.064458 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4bbd2d2_37a4_46d1_b487_1ece965b4a11.slice/crio-c9066bdab2582fbfdb47955d74a5c388d9d9a627253f5e2117d96566265f0ae8 WatchSource:0}: Error finding container c9066bdab2582fbfdb47955d74a5c388d9d9a627253f5e2117d96566265f0ae8: Status 404 returned error can't find the container with id c9066bdab2582fbfdb47955d74a5c388d9d9a627253f5e2117d96566265f0ae8 Dec 03 17:16:01 crc kubenswrapper[4758]: I1203 17:16:01.211185 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4bbd2d2-37a4-46d1-b487-1ece965b4a11","Type":"ContainerStarted","Data":"c9066bdab2582fbfdb47955d74a5c388d9d9a627253f5e2117d96566265f0ae8"} Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.072877 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5czb6"] Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.074401 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.080080 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.080631 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.080848 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qdkpn" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.080900 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6"] Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104594 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg224\" (UniqueName: \"kubernetes.io/projected/aac23ac3-72ea-49a7-9ded-6fabe28ff690-kube-api-access-rg224\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104647 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-log-ovn\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104671 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-ovn-controller-tls-certs\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104720 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104775 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104836 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-combined-ca-bundle\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.104862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run-ovn\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.140886 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-5mttb"] Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.142761 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.155845 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5mttb"] Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207004 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-scripts\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207062 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg224\" (UniqueName: \"kubernetes.io/projected/aac23ac3-72ea-49a7-9ded-6fabe28ff690-kube-api-access-rg224\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-log-ovn\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207110 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-ovn-controller-tls-certs\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207137 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-log\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207196 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nb2n\" (UniqueName: \"kubernetes.io/projected/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-kube-api-access-2nb2n\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207231 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207257 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-etc-ovs\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207295 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-run\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207396 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207454 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-combined-ca-bundle\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-lib\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.207513 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run-ovn\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.208309 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run-ovn\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.208904 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-log-ovn\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.213364 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.216383 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.255523 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-ovn-controller-tls-certs\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.255955 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-combined-ca-bundle\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.261972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg224\" (UniqueName: \"kubernetes.io/projected/aac23ac3-72ea-49a7-9ded-6fabe28ff690-kube-api-access-rg224\") pod \"ovn-controller-5czb6\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.310787 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-scripts\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.310871 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-log\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.310902 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nb2n\" (UniqueName: \"kubernetes.io/projected/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-kube-api-access-2nb2n\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.310927 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-etc-ovs\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.310958 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-run\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.311010 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-lib\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.311300 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-lib\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.314525 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-scripts\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.314651 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-log\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.315127 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-etc-ovs\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.315176 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-run\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.335770 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nb2n\" (UniqueName: \"kubernetes.io/projected/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-kube-api-access-2nb2n\") pod \"ovn-controller-ovs-5mttb\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.448084 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.501409 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.631543 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.633918 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.637321 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.638383 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.639287 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.639764 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.640148 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6nvqw" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.658372 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.847707 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848112 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848151 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848180 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-config\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848216 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848240 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848268 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pftgw\" (UniqueName: \"kubernetes.io/projected/d870c33d-67ba-4991-8624-4d0a28cb8542-kube-api-access-pftgw\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.848304 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949423 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949517 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949636 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-config\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949671 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949702 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.949730 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pftgw\" (UniqueName: \"kubernetes.io/projected/d870c33d-67ba-4991-8624-4d0a28cb8542-kube-api-access-pftgw\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.951428 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-config\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.951986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.952402 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.954477 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.968626 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.975020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.979168 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pftgw\" (UniqueName: \"kubernetes.io/projected/d870c33d-67ba-4991-8624-4d0a28cb8542-kube-api-access-pftgw\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:02 crc kubenswrapper[4758]: I1203 17:16:02.987221 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:03 crc kubenswrapper[4758]: I1203 17:16:03.005724 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:03 crc kubenswrapper[4758]: I1203 17:16:03.277450 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:03 crc kubenswrapper[4758]: I1203 17:16:03.560161 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6"] Dec 03 17:16:03 crc kubenswrapper[4758]: I1203 17:16:03.856000 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5mttb"] Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.561452 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-55pgs"] Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.566240 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.572826 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.581740 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-55pgs"] Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.700634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.700765 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovs-rundir\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.700801 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovn-rundir\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.700835 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2923bcc5-72ec-493c-ad10-9f171801ae4a-config\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.700853 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-combined-ca-bundle\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.700877 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pz2f\" (UniqueName: \"kubernetes.io/projected/2923bcc5-72ec-493c-ad10-9f171801ae4a-kube-api-access-8pz2f\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.802349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.803049 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovs-rundir\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.803207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovn-rundir\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.803319 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2923bcc5-72ec-493c-ad10-9f171801ae4a-config\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.803358 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-combined-ca-bundle\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.803414 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pz2f\" (UniqueName: \"kubernetes.io/projected/2923bcc5-72ec-493c-ad10-9f171801ae4a-kube-api-access-8pz2f\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.803627 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovn-rundir\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.804154 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovs-rundir\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.804159 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2923bcc5-72ec-493c-ad10-9f171801ae4a-config\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.810335 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.817377 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-combined-ca-bundle\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.827547 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pz2f\" (UniqueName: \"kubernetes.io/projected/2923bcc5-72ec-493c-ad10-9f171801ae4a-kube-api-access-8pz2f\") pod \"ovn-controller-metrics-55pgs\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:04 crc kubenswrapper[4758]: I1203 17:16:04.898986 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.713224 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.715877 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.719775 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.719837 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.720179 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-95l4r" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.720713 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.750074 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.887625 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.887725 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.887762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-config\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.887794 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwlhn\" (UniqueName: \"kubernetes.io/projected/7a370910-1957-4bd8-891b-7f1065c5e286-kube-api-access-rwlhn\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.888015 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.888239 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.888557 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.888819 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.995736 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.995828 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.995872 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-config\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.995897 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwlhn\" (UniqueName: \"kubernetes.io/projected/7a370910-1957-4bd8-891b-7f1065c5e286-kube-api-access-rwlhn\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.996009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.996133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.996200 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:05 crc kubenswrapper[4758]: I1203 17:16:05.996229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.000510 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.003621 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.004851 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.005251 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.005573 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-config\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.129933 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwlhn\" (UniqueName: \"kubernetes.io/projected/7a370910-1957-4bd8-891b-7f1065c5e286-kube-api-access-rwlhn\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.131992 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.135599 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.135807 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:06 crc kubenswrapper[4758]: I1203 17:16:06.346379 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:22 crc kubenswrapper[4758]: I1203 17:16:22.507830 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerStarted","Data":"2692fbc5fd698183cfebe16cf94c0a605e5e604f7a69a78096bf4fd2a8c9c8d0"} Dec 03 17:16:22 crc kubenswrapper[4758]: I1203 17:16:22.510753 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6" event={"ID":"aac23ac3-72ea-49a7-9ded-6fabe28ff690","Type":"ContainerStarted","Data":"5254ee401bfb1f603de824aaac045083787e69c235fdd2b1fd80e2eb768d747f"} Dec 03 17:16:23 crc kubenswrapper[4758]: E1203 17:16:23.419145 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 03 17:16:23 crc kubenswrapper[4758]: E1203 17:16:23.420216 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5cch5f4h646h55hf8h665h64bh667h647h75h67ch588h5d7hddh98h54dhc5hc6h545h556h85hc6h5d6h95h55fhb6h695h5b7hfbh55dh689h5f5q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9w2dg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(0754ea69-5007-4e51-bd22-0efd13954199): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:23 crc kubenswrapper[4758]: E1203 17:16:23.421607 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="0754ea69-5007-4e51-bd22-0efd13954199" Dec 03 17:16:23 crc kubenswrapper[4758]: E1203 17:16:23.523426 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="0754ea69-5007-4e51-bd22-0efd13954199" Dec 03 17:16:26 crc kubenswrapper[4758]: E1203 17:16:26.310990 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 03 17:16:26 crc kubenswrapper[4758]: E1203 17:16:26.311620 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ggl67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(80b1161a-3f70-4956-b13a-1ef693aa2e7c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:26 crc kubenswrapper[4758]: E1203 17:16:26.312931 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" Dec 03 17:16:26 crc kubenswrapper[4758]: E1203 17:16:26.550138 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.629009 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:650ba1d1fb09dfb2a96481f2ebae84fbae4b3474d057dd6d1569ce094dc41659: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-ovn-controller/blobs/sha256:650ba1d1fb09dfb2a96481f2ebae84fbae4b3474d057dd6d1569ce094dc41659\": context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.629940 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd9hd6h66fh54ch698hf6h656h9dh6ch594h588h558hbdh4h567hf8h57dh7bh5b8hcch65fh5dchbfh657hffh5b8h5f9h5b4h6dh5ddh5b6h86q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rg224,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-5czb6_openstack(aac23ac3-72ea-49a7-9ded-6fabe28ff690): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:650ba1d1fb09dfb2a96481f2ebae84fbae4b3474d057dd6d1569ce094dc41659: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-ovn-controller/blobs/sha256:650ba1d1fb09dfb2a96481f2ebae84fbae4b3474d057dd6d1569ce094dc41659\": context canceled" logger="UnhandledError" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.630321 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.630514 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8mjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(e518390e-b4c4-49bb-913f-ddfc331456ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.632481 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.632552 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:650ba1d1fb09dfb2a96481f2ebae84fbae4b3474d057dd6d1569ce094dc41659: Get \\\"https://quay.io/v2/podified-antelope-centos9/openstack-ovn-controller/blobs/sha256:650ba1d1fb09dfb2a96481f2ebae84fbae4b3474d057dd6d1569ce094dc41659\\\": context canceled\"" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.685933 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.686194 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ccv6m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(9c80de0e-2bee-4e3f-8edf-f4a235b01d76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.687824 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.717631 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.717937 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gk9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(070406cc-6d65-4fd0-81d7-264adead6bbd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.719402 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.798943 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.799393 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.799531 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" Dec 03 17:16:39 crc kubenswrapper[4758]: E1203 17:16:39.799624 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.281365 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.282104 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m8mns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-p9z46_openstack(cc544ec4-59b0-4acf-a808-db315c913f98): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.283369 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" podUID="cc544ec4-59b0-4acf-a808-db315c913f98" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.310205 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.310449 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzrvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-tj6n2_openstack(656169ac-4510-47b8-961c-ae8d2439af94): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.311751 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" podUID="656169ac-4510-47b8-961c-ae8d2439af94" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.318294 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.318577 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bt5km,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-c5v42_openstack(cdd798d4-66a3-4087-a7c8-bd1bbc101664): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.319973 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" podUID="cdd798d4-66a3-4087-a7c8-bd1bbc101664" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.779575 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" podUID="cc544ec4-59b0-4acf-a808-db315c913f98" Dec 03 17:16:41 crc kubenswrapper[4758]: E1203 17:16:41.783825 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" podUID="cdd798d4-66a3-4087-a7c8-bd1bbc101664" Dec 03 17:16:42 crc kubenswrapper[4758]: E1203 17:16:42.843323 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified" Dec 03 17:16:42 crc kubenswrapper[4758]: E1203 17:16:42.843627 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd9hd6h66fh54ch698hf6h656h9dh6ch594h588h558hbdh4h567hf8h57dh7bh5b8hcch65fh5dchbfh657hffh5b8h5f9h5b4h6dh5ddh5b6h86q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2nb2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-5mttb_openstack(3e30b3f9-d35f-43a0-9f06-25ebf09ccb34): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:42 crc kubenswrapper[4758]: E1203 17:16:42.844996 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" Dec 03 17:16:42 crc kubenswrapper[4758]: E1203 17:16:42.873914 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:16:42 crc kubenswrapper[4758]: E1203 17:16:42.874129 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4tzg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-5wcgm_openstack(8f96b7cb-0e42-4124-8ccb-f5b66d9ef992): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:16:42 crc kubenswrapper[4758]: E1203 17:16:42.875322 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" podUID="8f96b7cb-0e42-4124-8ccb-f5b66d9ef992" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.038573 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.166190 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzrvn\" (UniqueName: \"kubernetes.io/projected/656169ac-4510-47b8-961c-ae8d2439af94-kube-api-access-vzrvn\") pod \"656169ac-4510-47b8-961c-ae8d2439af94\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.166371 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-dns-svc\") pod \"656169ac-4510-47b8-961c-ae8d2439af94\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.166465 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-config\") pod \"656169ac-4510-47b8-961c-ae8d2439af94\" (UID: \"656169ac-4510-47b8-961c-ae8d2439af94\") " Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.167308 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-config" (OuterVolumeSpecName: "config") pod "656169ac-4510-47b8-961c-ae8d2439af94" (UID: "656169ac-4510-47b8-961c-ae8d2439af94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.167473 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "656169ac-4510-47b8-961c-ae8d2439af94" (UID: "656169ac-4510-47b8-961c-ae8d2439af94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.174871 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656169ac-4510-47b8-961c-ae8d2439af94-kube-api-access-vzrvn" (OuterVolumeSpecName: "kube-api-access-vzrvn") pod "656169ac-4510-47b8-961c-ae8d2439af94" (UID: "656169ac-4510-47b8-961c-ae8d2439af94"). InnerVolumeSpecName "kube-api-access-vzrvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.269055 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.269103 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656169ac-4510-47b8-961c-ae8d2439af94-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.269117 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzrvn\" (UniqueName: \"kubernetes.io/projected/656169ac-4510-47b8-961c-ae8d2439af94-kube-api-access-vzrvn\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.479508 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-55pgs"] Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.487307 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.797177 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" event={"ID":"656169ac-4510-47b8-961c-ae8d2439af94","Type":"ContainerDied","Data":"d51bb8ac07e6f17cb097dfb12dbea6c8c80d9d6df0772fa127aa85dd37e48445"} Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.797749 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tj6n2" Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.800628 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-55pgs" event={"ID":"2923bcc5-72ec-493c-ad10-9f171801ae4a","Type":"ContainerStarted","Data":"a094f01db20cff1eb297edaa1f2f22fb6e131a443fa835e281e32bedc079b363"} Dec 03 17:16:43 crc kubenswrapper[4758]: E1203 17:16:43.802726 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified\\\"\"" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" Dec 03 17:16:43 crc kubenswrapper[4758]: W1203 17:16:43.821153 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a370910_1957_4bd8_891b_7f1065c5e286.slice/crio-67b7edf337a173465cbd23ee0f7c74958897500d7f7dff5d3ac774d3025af670 WatchSource:0}: Error finding container 67b7edf337a173465cbd23ee0f7c74958897500d7f7dff5d3ac774d3025af670: Status 404 returned error can't find the container with id 67b7edf337a173465cbd23ee0f7c74958897500d7f7dff5d3ac774d3025af670 Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.905025 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tj6n2"] Dec 03 17:16:43 crc kubenswrapper[4758]: I1203 17:16:43.912352 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tj6n2"] Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.237444 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.297936 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-config\") pod \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.298021 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tzg7\" (UniqueName: \"kubernetes.io/projected/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-kube-api-access-4tzg7\") pod \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\" (UID: \"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992\") " Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.298759 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-config" (OuterVolumeSpecName: "config") pod "8f96b7cb-0e42-4124-8ccb-f5b66d9ef992" (UID: "8f96b7cb-0e42-4124-8ccb-f5b66d9ef992"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.303896 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-kube-api-access-4tzg7" (OuterVolumeSpecName: "kube-api-access-4tzg7") pod "8f96b7cb-0e42-4124-8ccb-f5b66d9ef992" (UID: "8f96b7cb-0e42-4124-8ccb-f5b66d9ef992"). InnerVolumeSpecName "kube-api-access-4tzg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.400213 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.400877 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tzg7\" (UniqueName: \"kubernetes.io/projected/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992-kube-api-access-4tzg7\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.444383 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:16:44 crc kubenswrapper[4758]: W1203 17:16:44.458869 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd870c33d_67ba_4991_8624_4d0a28cb8542.slice/crio-d1d2ed1558de329762a10a5d07018fcffb9a01af10ac51b61adc356e247adcc2 WatchSource:0}: Error finding container d1d2ed1558de329762a10a5d07018fcffb9a01af10ac51b61adc356e247adcc2: Status 404 returned error can't find the container with id d1d2ed1558de329762a10a5d07018fcffb9a01af10ac51b61adc356e247adcc2 Dec 03 17:16:44 crc kubenswrapper[4758]: E1203 17:16:44.706141 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6: Get \"https://registry.k8s.io/v2/kube-state-metrics/kube-state-metrics/blobs/sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6\": context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 03 17:16:44 crc kubenswrapper[4758]: E1203 17:16:44.706235 4758 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6: Get \"https://registry.k8s.io/v2/kube-state-metrics/kube-state-metrics/blobs/sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6\": context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 03 17:16:44 crc kubenswrapper[4758]: E1203 17:16:44.706564 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f8njx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(d4bbd2d2-37a4-46d1-b487-1ece965b4a11): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6: Get \"https://registry.k8s.io/v2/kube-state-metrics/kube-state-metrics/blobs/sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6\": context canceled" logger="UnhandledError" Dec 03 17:16:44 crc kubenswrapper[4758]: E1203 17:16:44.707830 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6: Get \\\"https://registry.k8s.io/v2/kube-state-metrics/kube-state-metrics/blobs/sha256:da7816fa955ea24533c388143c78804c28682eef99b4ee3723b548c70148bba6\\\": context canceled\"" pod="openstack/kube-state-metrics-0" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.838362 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d870c33d-67ba-4991-8624-4d0a28cb8542","Type":"ContainerStarted","Data":"d1d2ed1558de329762a10a5d07018fcffb9a01af10ac51b61adc356e247adcc2"} Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.864071 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7a370910-1957-4bd8-891b-7f1065c5e286","Type":"ContainerStarted","Data":"67b7edf337a173465cbd23ee0f7c74958897500d7f7dff5d3ac774d3025af670"} Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.889141 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" event={"ID":"8f96b7cb-0e42-4124-8ccb-f5b66d9ef992","Type":"ContainerDied","Data":"01e16815e497d58d1a41a615a72a7d0adc7710b037aa4b1a795f513bebeed9d5"} Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.889321 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5wcgm" Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.906952 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0754ea69-5007-4e51-bd22-0efd13954199","Type":"ContainerStarted","Data":"2071a07b9d8bd2fffbcfc63e0cd6e8e249aa95143b551ab078f950658b27507b"} Dec 03 17:16:44 crc kubenswrapper[4758]: I1203 17:16:44.907818 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 17:16:44 crc kubenswrapper[4758]: E1203 17:16:44.908499 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.021936 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5wcgm"] Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.030373 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5wcgm"] Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.040329 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.729651901 podStartE2EDuration="49.040299363s" podCreationTimestamp="2025-12-03 17:15:56 +0000 UTC" firstStartedPulling="2025-12-03 17:15:58.219207623 +0000 UTC m=+1213.420584484" lastFinishedPulling="2025-12-03 17:16:43.529855085 +0000 UTC m=+1258.731231946" observedRunningTime="2025-12-03 17:16:45.019253633 +0000 UTC m=+1260.220630514" watchObservedRunningTime="2025-12-03 17:16:45.040299363 +0000 UTC m=+1260.241676224" Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.128227 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656169ac-4510-47b8-961c-ae8d2439af94" path="/var/lib/kubelet/pods/656169ac-4510-47b8-961c-ae8d2439af94/volumes" Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.128744 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f96b7cb-0e42-4124-8ccb-f5b66d9ef992" path="/var/lib/kubelet/pods/8f96b7cb-0e42-4124-8ccb-f5b66d9ef992/volumes" Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.918608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7a370910-1957-4bd8-891b-7f1065c5e286","Type":"ContainerStarted","Data":"a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098"} Dec 03 17:16:45 crc kubenswrapper[4758]: I1203 17:16:45.922608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"80b1161a-3f70-4956-b13a-1ef693aa2e7c","Type":"ContainerStarted","Data":"664e30dda3d80d735a4b98f34b975e32bed09a6c3a4d4ec313abc561a54aae69"} Dec 03 17:16:47 crc kubenswrapper[4758]: I1203 17:16:47.943960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7a370910-1957-4bd8-891b-7f1065c5e286","Type":"ContainerStarted","Data":"da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a"} Dec 03 17:16:47 crc kubenswrapper[4758]: I1203 17:16:47.946842 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-55pgs" event={"ID":"2923bcc5-72ec-493c-ad10-9f171801ae4a","Type":"ContainerStarted","Data":"5877538da9e694dfac7b157eb99747894a3276ae973e31c6d9517ea26fd87fd7"} Dec 03 17:16:47 crc kubenswrapper[4758]: I1203 17:16:47.948356 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d870c33d-67ba-4991-8624-4d0a28cb8542","Type":"ContainerStarted","Data":"66ede82dc9e0fb893690735dd61d3b58c7e3794bb7c61e6dce24226c9e1b2e55"} Dec 03 17:16:47 crc kubenswrapper[4758]: I1203 17:16:47.979108 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=40.127088167 podStartE2EDuration="43.979085455s" podCreationTimestamp="2025-12-03 17:16:04 +0000 UTC" firstStartedPulling="2025-12-03 17:16:43.836504168 +0000 UTC m=+1259.037881029" lastFinishedPulling="2025-12-03 17:16:47.688501456 +0000 UTC m=+1262.889878317" observedRunningTime="2025-12-03 17:16:47.97076875 +0000 UTC m=+1263.172145631" watchObservedRunningTime="2025-12-03 17:16:47.979085455 +0000 UTC m=+1263.180462316" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.000273 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-55pgs" podStartSLOduration=39.879708817 podStartE2EDuration="44.000250128s" podCreationTimestamp="2025-12-03 17:16:04 +0000 UTC" firstStartedPulling="2025-12-03 17:16:43.582074659 +0000 UTC m=+1258.783451520" lastFinishedPulling="2025-12-03 17:16:47.70261597 +0000 UTC m=+1262.903992831" observedRunningTime="2025-12-03 17:16:47.992989121 +0000 UTC m=+1263.194365992" watchObservedRunningTime="2025-12-03 17:16:48.000250128 +0000 UTC m=+1263.201626989" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.347750 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.391673 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.400801 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5v42"] Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.496010 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fjvmb"] Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.525472 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.532254 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.548087 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fjvmb"] Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.616625 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn64z\" (UniqueName: \"kubernetes.io/projected/91a9316f-a950-4651-98c3-4dd2b0d247bc-kube-api-access-nn64z\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.617943 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.618003 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-config\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.618078 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.678699 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p9z46"] Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.721038 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.721132 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-config\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.721187 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.721236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn64z\" (UniqueName: \"kubernetes.io/projected/91a9316f-a950-4651-98c3-4dd2b0d247bc-kube-api-access-nn64z\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.722638 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.723188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-config\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.738141 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.815233 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn64z\" (UniqueName: \"kubernetes.io/projected/91a9316f-a950-4651-98c3-4dd2b0d247bc-kube-api-access-nn64z\") pod \"dnsmasq-dns-5bf47b49b7-fjvmb\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.820617 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-xwrl6"] Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.822498 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.828413 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.901014 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.926698 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.926766 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-dns-svc\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.926791 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsbb8\" (UniqueName: \"kubernetes.io/projected/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-kube-api-access-rsbb8\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.926854 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-config\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.926965 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:48 crc kubenswrapper[4758]: I1203 17:16:48.953915 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xwrl6"] Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.032965 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.033034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-dns-svc\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.033056 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsbb8\" (UniqueName: \"kubernetes.io/projected/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-kube-api-access-rsbb8\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.033104 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-config\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.033189 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.034177 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.034751 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.035252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-dns-svc\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.036415 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-config\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.079829 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsbb8\" (UniqueName: \"kubernetes.io/projected/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-kube-api-access-rsbb8\") pod \"dnsmasq-dns-8554648995-xwrl6\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.088265 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d870c33d-67ba-4991-8624-4d0a28cb8542","Type":"ContainerStarted","Data":"ec3db493149a77626469cdaa439ac0d0670962001d5407b77c4d6b653c833931"} Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.088836 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.119176 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.143380 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=44.633497674 podStartE2EDuration="48.14335228s" podCreationTimestamp="2025-12-03 17:16:01 +0000 UTC" firstStartedPulling="2025-12-03 17:16:44.463639878 +0000 UTC m=+1259.665016739" lastFinishedPulling="2025-12-03 17:16:47.973494484 +0000 UTC m=+1263.174871345" observedRunningTime="2025-12-03 17:16:49.136497514 +0000 UTC m=+1264.337874375" watchObservedRunningTime="2025-12-03 17:16:49.14335228 +0000 UTC m=+1264.344729141" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.197885 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.236132 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-config\") pod \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.236357 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-dns-svc\") pod \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.236397 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt5km\" (UniqueName: \"kubernetes.io/projected/cdd798d4-66a3-4087-a7c8-bd1bbc101664-kube-api-access-bt5km\") pod \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\" (UID: \"cdd798d4-66a3-4087-a7c8-bd1bbc101664\") " Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.238844 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-config" (OuterVolumeSpecName: "config") pod "cdd798d4-66a3-4087-a7c8-bd1bbc101664" (UID: "cdd798d4-66a3-4087-a7c8-bd1bbc101664"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.239287 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdd798d4-66a3-4087-a7c8-bd1bbc101664" (UID: "cdd798d4-66a3-4087-a7c8-bd1bbc101664"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.259665 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd798d4-66a3-4087-a7c8-bd1bbc101664-kube-api-access-bt5km" (OuterVolumeSpecName: "kube-api-access-bt5km") pod "cdd798d4-66a3-4087-a7c8-bd1bbc101664" (UID: "cdd798d4-66a3-4087-a7c8-bd1bbc101664"). InnerVolumeSpecName "kube-api-access-bt5km". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.345027 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.345472 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt5km\" (UniqueName: \"kubernetes.io/projected/cdd798d4-66a3-4087-a7c8-bd1bbc101664-kube-api-access-bt5km\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.345484 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdd798d4-66a3-4087-a7c8-bd1bbc101664-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.442744 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.551073 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8mns\" (UniqueName: \"kubernetes.io/projected/cc544ec4-59b0-4acf-a808-db315c913f98-kube-api-access-m8mns\") pod \"cc544ec4-59b0-4acf-a808-db315c913f98\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.551226 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-dns-svc\") pod \"cc544ec4-59b0-4acf-a808-db315c913f98\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.551328 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-config\") pod \"cc544ec4-59b0-4acf-a808-db315c913f98\" (UID: \"cc544ec4-59b0-4acf-a808-db315c913f98\") " Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.551912 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc544ec4-59b0-4acf-a808-db315c913f98" (UID: "cc544ec4-59b0-4acf-a808-db315c913f98"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.552513 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-config" (OuterVolumeSpecName: "config") pod "cc544ec4-59b0-4acf-a808-db315c913f98" (UID: "cc544ec4-59b0-4acf-a808-db315c913f98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.553469 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.553499 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc544ec4-59b0-4acf-a808-db315c913f98-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.555916 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc544ec4-59b0-4acf-a808-db315c913f98-kube-api-access-m8mns" (OuterVolumeSpecName: "kube-api-access-m8mns") pod "cc544ec4-59b0-4acf-a808-db315c913f98" (UID: "cc544ec4-59b0-4acf-a808-db315c913f98"). InnerVolumeSpecName "kube-api-access-m8mns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.655944 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8mns\" (UniqueName: \"kubernetes.io/projected/cc544ec4-59b0-4acf-a808-db315c913f98-kube-api-access-m8mns\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.669249 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fjvmb"] Dec 03 17:16:49 crc kubenswrapper[4758]: W1203 17:16:49.671862 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91a9316f_a950_4651_98c3_4dd2b0d247bc.slice/crio-2a757bdd604f28b2f64557afd0f01c586dd87d685658f4f12d15841c149fc725 WatchSource:0}: Error finding container 2a757bdd604f28b2f64557afd0f01c586dd87d685658f4f12d15841c149fc725: Status 404 returned error can't find the container with id 2a757bdd604f28b2f64557afd0f01c586dd87d685658f4f12d15841c149fc725 Dec 03 17:16:49 crc kubenswrapper[4758]: I1203 17:16:49.763208 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xwrl6"] Dec 03 17:16:49 crc kubenswrapper[4758]: W1203 17:16:49.763461 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda61c673f_5b2a_45b7_96d9_b5cc4c5749b1.slice/crio-27352ef3d9845af949f946fd24eb9d9c07e6c2b1cdc501333738daff088e67f6 WatchSource:0}: Error finding container 27352ef3d9845af949f946fd24eb9d9c07e6c2b1cdc501333738daff088e67f6: Status 404 returned error can't find the container with id 27352ef3d9845af949f946fd24eb9d9c07e6c2b1cdc501333738daff088e67f6 Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.096486 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.096470 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p9z46" event={"ID":"cc544ec4-59b0-4acf-a808-db315c913f98","Type":"ContainerDied","Data":"47e6f83d9af8d95ec8e103fee1b1ea568c43b40f6d33aa6c03578e4f424b3e6f"} Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.098474 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.098503 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-c5v42" event={"ID":"cdd798d4-66a3-4087-a7c8-bd1bbc101664","Type":"ContainerDied","Data":"c0ecb0a2b67c7ef56dc337a6ec25907155ad8ffb9b55b2f5e4471a539232dd18"} Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.100284 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" event={"ID":"91a9316f-a950-4651-98c3-4dd2b0d247bc","Type":"ContainerStarted","Data":"2a757bdd604f28b2f64557afd0f01c586dd87d685658f4f12d15841c149fc725"} Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.102491 4758 generic.go:334] "Generic (PLEG): container finished" podID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerID="664e30dda3d80d735a4b98f34b975e32bed09a6c3a4d4ec313abc561a54aae69" exitCode=0 Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.102588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"80b1161a-3f70-4956-b13a-1ef693aa2e7c","Type":"ContainerDied","Data":"664e30dda3d80d735a4b98f34b975e32bed09a6c3a4d4ec313abc561a54aae69"} Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.107316 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xwrl6" event={"ID":"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1","Type":"ContainerStarted","Data":"27352ef3d9845af949f946fd24eb9d9c07e6c2b1cdc501333738daff088e67f6"} Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.177651 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.209946 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5v42"] Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.228320 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5v42"] Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.260782 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p9z46"] Dec 03 17:16:50 crc kubenswrapper[4758]: I1203 17:16:50.266330 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p9z46"] Dec 03 17:16:51 crc kubenswrapper[4758]: I1203 17:16:51.125292 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc544ec4-59b0-4acf-a808-db315c913f98" path="/var/lib/kubelet/pods/cc544ec4-59b0-4acf-a808-db315c913f98/volumes" Dec 03 17:16:51 crc kubenswrapper[4758]: I1203 17:16:51.126208 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd798d4-66a3-4087-a7c8-bd1bbc101664" path="/var/lib/kubelet/pods/cdd798d4-66a3-4087-a7c8-bd1bbc101664/volumes" Dec 03 17:16:51 crc kubenswrapper[4758]: I1203 17:16:51.278430 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:51 crc kubenswrapper[4758]: I1203 17:16:51.323961 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.132750 4758 generic.go:334] "Generic (PLEG): container finished" podID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerID="fe0431b466f23c5236045a7abb2108a2f3b4390cfee17ffd736b71cb61050e03" exitCode=0 Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.132846 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xwrl6" event={"ID":"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1","Type":"ContainerDied","Data":"fe0431b466f23c5236045a7abb2108a2f3b4390cfee17ffd736b71cb61050e03"} Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.137391 4758 generic.go:334] "Generic (PLEG): container finished" podID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerID="ea25920413203b692677795e309ef5a10915921afc7f2994986ca677e68ced49" exitCode=0 Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.137533 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" event={"ID":"91a9316f-a950-4651-98c3-4dd2b0d247bc","Type":"ContainerDied","Data":"ea25920413203b692677795e309ef5a10915921afc7f2994986ca677e68ced49"} Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.141764 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"80b1161a-3f70-4956-b13a-1ef693aa2e7c","Type":"ContainerStarted","Data":"d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42"} Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.141986 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.193728 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.572101699 podStartE2EDuration="57.193690532s" podCreationTimestamp="2025-12-03 17:15:55 +0000 UTC" firstStartedPulling="2025-12-03 17:15:57.823671498 +0000 UTC m=+1213.025048359" lastFinishedPulling="2025-12-03 17:16:44.445260331 +0000 UTC m=+1259.646637192" observedRunningTime="2025-12-03 17:16:52.184308058 +0000 UTC m=+1267.385684919" watchObservedRunningTime="2025-12-03 17:16:52.193690532 +0000 UTC m=+1267.395067393" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.228020 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.302302 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.581657 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.584546 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.590571 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.590638 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.591094 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pm77p" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.591098 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.610714 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717437 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-config\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717509 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717544 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbmxq\" (UniqueName: \"kubernetes.io/projected/daf64139-0696-484a-870f-023898fbeb08-kube-api-access-nbmxq\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717568 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-scripts\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717619 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717665 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daf64139-0696-484a-870f-023898fbeb08-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.717717 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.819840 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-config\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.819907 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.819934 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbmxq\" (UniqueName: \"kubernetes.io/projected/daf64139-0696-484a-870f-023898fbeb08-kube-api-access-nbmxq\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.819957 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-scripts\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.820001 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.820034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daf64139-0696-484a-870f-023898fbeb08-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.820051 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.821309 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daf64139-0696-484a-870f-023898fbeb08-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.821354 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-config\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.822844 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-scripts\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.827575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.827661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.829517 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.842670 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbmxq\" (UniqueName: \"kubernetes.io/projected/daf64139-0696-484a-870f-023898fbeb08-kube-api-access-nbmxq\") pod \"ovn-northd-0\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " pod="openstack/ovn-northd-0" Dec 03 17:16:52 crc kubenswrapper[4758]: I1203 17:16:52.939430 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.180039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xwrl6" event={"ID":"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1","Type":"ContainerStarted","Data":"0808c7b33148d49a50727b199a344d2d5e5ae74f8fc8755e99598b3aaaffe6b6"} Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.180503 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.185848 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" event={"ID":"91a9316f-a950-4651-98c3-4dd2b0d247bc","Type":"ContainerStarted","Data":"acd55fb05433cbe0f69bf17d955d194bd04fcebf58b4093178515fbf1c5a19e2"} Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.185935 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.207415 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-xwrl6" podStartSLOduration=3.433840629 podStartE2EDuration="5.207377089s" podCreationTimestamp="2025-12-03 17:16:48 +0000 UTC" firstStartedPulling="2025-12-03 17:16:49.770866101 +0000 UTC m=+1264.972242962" lastFinishedPulling="2025-12-03 17:16:51.544402561 +0000 UTC m=+1266.745779422" observedRunningTime="2025-12-03 17:16:53.203191016 +0000 UTC m=+1268.404567877" watchObservedRunningTime="2025-12-03 17:16:53.207377089 +0000 UTC m=+1268.408753960" Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.239474 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" podStartSLOduration=3.640523305 podStartE2EDuration="5.239454988s" podCreationTimestamp="2025-12-03 17:16:48 +0000 UTC" firstStartedPulling="2025-12-03 17:16:49.674290306 +0000 UTC m=+1264.875667167" lastFinishedPulling="2025-12-03 17:16:51.273221989 +0000 UTC m=+1266.474598850" observedRunningTime="2025-12-03 17:16:53.238068101 +0000 UTC m=+1268.439444982" watchObservedRunningTime="2025-12-03 17:16:53.239454988 +0000 UTC m=+1268.440831859" Dec 03 17:16:53 crc kubenswrapper[4758]: I1203 17:16:53.447996 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:16:53 crc kubenswrapper[4758]: W1203 17:16:53.456662 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf64139_0696_484a_870f_023898fbeb08.slice/crio-f1c7d1ad6d07d74047a1d1601a9af41bcc9dba404849ff99f0561d54cb0315b9 WatchSource:0}: Error finding container f1c7d1ad6d07d74047a1d1601a9af41bcc9dba404849ff99f0561d54cb0315b9: Status 404 returned error can't find the container with id f1c7d1ad6d07d74047a1d1601a9af41bcc9dba404849ff99f0561d54cb0315b9 Dec 03 17:16:54 crc kubenswrapper[4758]: I1203 17:16:54.218137 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"daf64139-0696-484a-870f-023898fbeb08","Type":"ContainerStarted","Data":"f1c7d1ad6d07d74047a1d1601a9af41bcc9dba404849ff99f0561d54cb0315b9"} Dec 03 17:16:55 crc kubenswrapper[4758]: I1203 17:16:55.228786 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9c80de0e-2bee-4e3f-8edf-f4a235b01d76","Type":"ContainerStarted","Data":"ef2a638cba2c7b4383584dce02eb1679141386353eb7c10bd4bd9af144b23c05"} Dec 03 17:16:56 crc kubenswrapper[4758]: I1203 17:16:56.248222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e518390e-b4c4-49bb-913f-ddfc331456ab","Type":"ContainerStarted","Data":"085ad6a4f1116fe1014b94a49396f4beca5e6b5a210c06edc52bb5e34006f632"} Dec 03 17:16:56 crc kubenswrapper[4758]: I1203 17:16:56.251633 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"070406cc-6d65-4fd0-81d7-264adead6bbd","Type":"ContainerStarted","Data":"6de2813edf252cbe7e4d319d670dd48259def0afd5b8b085c56a64c56604dd50"} Dec 03 17:16:57 crc kubenswrapper[4758]: I1203 17:16:57.102782 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 17:16:57 crc kubenswrapper[4758]: I1203 17:16:57.103263 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 17:16:57 crc kubenswrapper[4758]: I1203 17:16:57.267579 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"daf64139-0696-484a-870f-023898fbeb08","Type":"ContainerStarted","Data":"c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7"} Dec 03 17:16:58 crc kubenswrapper[4758]: I1203 17:16:58.903903 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.199935 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.270436 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fjvmb"] Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.292244 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="dnsmasq-dns" containerID="cri-o://acd55fb05433cbe0f69bf17d955d194bd04fcebf58b4093178515fbf1c5a19e2" gracePeriod=10 Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.530662 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-zpczk"] Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.532966 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.559056 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-zpczk"] Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.678133 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-config\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.678194 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.678248 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.678300 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2twf4\" (UniqueName: \"kubernetes.io/projected/566c8d53-04a9-4937-bdab-4cac68a6ca49-kube-api-access-2twf4\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.678429 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.780462 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-config\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.780545 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.780626 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.780732 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2twf4\" (UniqueName: \"kubernetes.io/projected/566c8d53-04a9-4937-bdab-4cac68a6ca49-kube-api-access-2twf4\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.780795 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.782138 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.783034 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-config\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.783137 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.783783 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.809081 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2twf4\" (UniqueName: \"kubernetes.io/projected/566c8d53-04a9-4937-bdab-4cac68a6ca49-kube-api-access-2twf4\") pod \"dnsmasq-dns-b8fbc5445-zpczk\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:16:59 crc kubenswrapper[4758]: I1203 17:16:59.864839 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.387565 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-zpczk"] Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.677374 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.685289 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.690023 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.690921 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-zptr6" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.691398 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.697103 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.699756 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.814219 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-lock\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.814583 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.814784 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-cache\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.815026 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vvcx\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-kube-api-access-2vvcx\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.815458 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.917426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.917879 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-lock\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.918000 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: E1203 17:17:00.917786 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:17:00 crc kubenswrapper[4758]: E1203 17:17:00.918155 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.918256 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-cache\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: E1203 17:17:00.918339 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift podName:946a9b5f-7abf-4b3e-bc51-08a45e73ff30 nodeName:}" failed. No retries permitted until 2025-12-03 17:17:01.418292215 +0000 UTC m=+1276.619669256 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift") pod "swift-storage-0" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30") : configmap "swift-ring-files" not found Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.918400 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vvcx\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-kube-api-access-2vvcx\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.918449 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-lock\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.918865 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.919116 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-cache\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.943979 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:00 crc kubenswrapper[4758]: I1203 17:17:00.946176 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vvcx\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-kube-api-access-2vvcx\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:01 crc kubenswrapper[4758]: I1203 17:17:01.324267 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" event={"ID":"566c8d53-04a9-4937-bdab-4cac68a6ca49","Type":"ContainerStarted","Data":"1cf0016e1f0b774f19cba679fc7b3cb41009855c5aea535c21e98ca34e4d9bf9"} Dec 03 17:17:01 crc kubenswrapper[4758]: I1203 17:17:01.328508 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"daf64139-0696-484a-870f-023898fbeb08","Type":"ContainerStarted","Data":"43d7c9ec7f4e1c689ef29fc35a5e148631fe6789db20d2476a17525c48bb0c03"} Dec 03 17:17:01 crc kubenswrapper[4758]: I1203 17:17:01.427886 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:01 crc kubenswrapper[4758]: E1203 17:17:01.428191 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:17:01 crc kubenswrapper[4758]: E1203 17:17:01.428231 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:17:01 crc kubenswrapper[4758]: E1203 17:17:01.428310 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift podName:946a9b5f-7abf-4b3e-bc51-08a45e73ff30 nodeName:}" failed. No retries permitted until 2025-12-03 17:17:02.428288415 +0000 UTC m=+1277.629665276 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift") pod "swift-storage-0" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30") : configmap "swift-ring-files" not found Dec 03 17:17:02 crc kubenswrapper[4758]: I1203 17:17:02.447368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:02 crc kubenswrapper[4758]: E1203 17:17:02.448385 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:17:02 crc kubenswrapper[4758]: E1203 17:17:02.448472 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:17:02 crc kubenswrapper[4758]: E1203 17:17:02.448654 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift podName:946a9b5f-7abf-4b3e-bc51-08a45e73ff30 nodeName:}" failed. No retries permitted until 2025-12-03 17:17:04.448638543 +0000 UTC m=+1279.650015394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift") pod "swift-storage-0" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30") : configmap "swift-ring-files" not found Dec 03 17:17:03 crc kubenswrapper[4758]: E1203 17:17:03.729323 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566c8d53_04a9_4937_bdab_4cac68a6ca49.slice/crio-50338516a62f4c77d97174e64b721bc57c9c6d94dadca1ad11e722975542178f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:17:03 crc kubenswrapper[4758]: I1203 17:17:03.867884 4758 generic.go:334] "Generic (PLEG): container finished" podID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerID="acd55fb05433cbe0f69bf17d955d194bd04fcebf58b4093178515fbf1c5a19e2" exitCode=0 Dec 03 17:17:03 crc kubenswrapper[4758]: I1203 17:17:03.867951 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" event={"ID":"91a9316f-a950-4651-98c3-4dd2b0d247bc","Type":"ContainerDied","Data":"acd55fb05433cbe0f69bf17d955d194bd04fcebf58b4093178515fbf1c5a19e2"} Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.398061 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.494670 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-ovsdbserver-nb\") pod \"91a9316f-a950-4651-98c3-4dd2b0d247bc\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.494808 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-dns-svc\") pod \"91a9316f-a950-4651-98c3-4dd2b0d247bc\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.494908 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn64z\" (UniqueName: \"kubernetes.io/projected/91a9316f-a950-4651-98c3-4dd2b0d247bc-kube-api-access-nn64z\") pod \"91a9316f-a950-4651-98c3-4dd2b0d247bc\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.495094 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-config\") pod \"91a9316f-a950-4651-98c3-4dd2b0d247bc\" (UID: \"91a9316f-a950-4651-98c3-4dd2b0d247bc\") " Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.495404 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:04 crc kubenswrapper[4758]: E1203 17:17:04.495645 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:17:04 crc kubenswrapper[4758]: E1203 17:17:04.495923 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:17:04 crc kubenswrapper[4758]: E1203 17:17:04.496012 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift podName:946a9b5f-7abf-4b3e-bc51-08a45e73ff30 nodeName:}" failed. No retries permitted until 2025-12-03 17:17:08.495983179 +0000 UTC m=+1283.697360040 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift") pod "swift-storage-0" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30") : configmap "swift-ring-files" not found Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.515001 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91a9316f-a950-4651-98c3-4dd2b0d247bc-kube-api-access-nn64z" (OuterVolumeSpecName: "kube-api-access-nn64z") pod "91a9316f-a950-4651-98c3-4dd2b0d247bc" (UID: "91a9316f-a950-4651-98c3-4dd2b0d247bc"). InnerVolumeSpecName "kube-api-access-nn64z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.543141 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "91a9316f-a950-4651-98c3-4dd2b0d247bc" (UID: "91a9316f-a950-4651-98c3-4dd2b0d247bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.546338 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "91a9316f-a950-4651-98c3-4dd2b0d247bc" (UID: "91a9316f-a950-4651-98c3-4dd2b0d247bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.563824 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-config" (OuterVolumeSpecName: "config") pod "91a9316f-a950-4651-98c3-4dd2b0d247bc" (UID: "91a9316f-a950-4651-98c3-4dd2b0d247bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.598136 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.598191 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.598204 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91a9316f-a950-4651-98c3-4dd2b0d247bc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.598218 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn64z\" (UniqueName: \"kubernetes.io/projected/91a9316f-a950-4651-98c3-4dd2b0d247bc-kube-api-access-nn64z\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.634265 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-48jkw"] Dec 03 17:17:04 crc kubenswrapper[4758]: E1203 17:17:04.634800 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="dnsmasq-dns" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.634823 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="dnsmasq-dns" Dec 03 17:17:04 crc kubenswrapper[4758]: E1203 17:17:04.634865 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="init" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.634873 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="init" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.635085 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="dnsmasq-dns" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.635839 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.643200 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.643663 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.643923 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.656812 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-48jkw"] Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.804885 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-ring-data-devices\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.806347 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-dispersionconf\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.806477 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1bc708e-493d-4597-9a37-501be2e8641a-etc-swift\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.806920 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-swiftconf\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.807229 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-scripts\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.807314 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4v9t\" (UniqueName: \"kubernetes.io/projected/e1bc708e-493d-4597-9a37-501be2e8641a-kube-api-access-j4v9t\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.807397 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-combined-ca-bundle\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.880331 4758 generic.go:334] "Generic (PLEG): container finished" podID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerID="50338516a62f4c77d97174e64b721bc57c9c6d94dadca1ad11e722975542178f" exitCode=0 Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.880450 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" event={"ID":"566c8d53-04a9-4937-bdab-4cac68a6ca49","Type":"ContainerDied","Data":"50338516a62f4c77d97174e64b721bc57c9c6d94dadca1ad11e722975542178f"} Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.883600 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" event={"ID":"91a9316f-a950-4651-98c3-4dd2b0d247bc","Type":"ContainerDied","Data":"2a757bdd604f28b2f64557afd0f01c586dd87d685658f4f12d15841c149fc725"} Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.883650 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.883743 4758 scope.go:117] "RemoveContainer" containerID="acd55fb05433cbe0f69bf17d955d194bd04fcebf58b4093178515fbf1c5a19e2" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.883814 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-ring-data-devices\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909397 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-dispersionconf\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1bc708e-493d-4597-9a37-501be2e8641a-etc-swift\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-swiftconf\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909618 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-scripts\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909646 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4v9t\" (UniqueName: \"kubernetes.io/projected/e1bc708e-493d-4597-9a37-501be2e8641a-kube-api-access-j4v9t\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.909724 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-combined-ca-bundle\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.915833 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1bc708e-493d-4597-9a37-501be2e8641a-etc-swift\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.916872 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-scripts\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.921697 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-ring-data-devices\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.926842 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-combined-ca-bundle\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.930265 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-swiftconf\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.936527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-dispersionconf\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.951135 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fjvmb"] Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.959518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4v9t\" (UniqueName: \"kubernetes.io/projected/e1bc708e-493d-4597-9a37-501be2e8641a-kube-api-access-j4v9t\") pod \"swift-ring-rebalance-48jkw\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.960866 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-fjvmb"] Dec 03 17:17:04 crc kubenswrapper[4758]: I1203 17:17:04.978255 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=10.436381422 podStartE2EDuration="12.978218896s" podCreationTimestamp="2025-12-03 17:16:52 +0000 UTC" firstStartedPulling="2025-12-03 17:16:53.463150385 +0000 UTC m=+1268.664527246" lastFinishedPulling="2025-12-03 17:16:56.004987859 +0000 UTC m=+1271.206364720" observedRunningTime="2025-12-03 17:17:04.967805494 +0000 UTC m=+1280.169182355" watchObservedRunningTime="2025-12-03 17:17:04.978218896 +0000 UTC m=+1280.179595767" Dec 03 17:17:05 crc kubenswrapper[4758]: I1203 17:17:05.004173 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:05 crc kubenswrapper[4758]: I1203 17:17:05.154302 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" path="/var/lib/kubelet/pods/91a9316f-a950-4651-98c3-4dd2b0d247bc/volumes" Dec 03 17:17:05 crc kubenswrapper[4758]: I1203 17:17:05.577031 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 17:17:05 crc kubenswrapper[4758]: I1203 17:17:05.668559 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 17:17:05 crc kubenswrapper[4758]: I1203 17:17:05.891355 4758 scope.go:117] "RemoveContainer" containerID="ea25920413203b692677795e309ef5a10915921afc7f2994986ca677e68ced49" Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.431368 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-48jkw"] Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.947427 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6" event={"ID":"aac23ac3-72ea-49a7-9ded-6fabe28ff690","Type":"ContainerStarted","Data":"12bf46a944f70cdab9ad1c78427dc024a221e2ca969030997c39eb4adbcaa629"} Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.949311 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-5czb6" Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.952655 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4bbd2d2-37a4-46d1-b487-1ece965b4a11","Type":"ContainerStarted","Data":"3e857659593ca56eeb3938a5a54b94b2417aab1032229a7c9c9ccbb10bc308db"} Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.953002 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.957181 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerStarted","Data":"883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74"} Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.960474 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" event={"ID":"566c8d53-04a9-4937-bdab-4cac68a6ca49","Type":"ContainerStarted","Data":"e2b441905ca3a543bb1a2a712ed67ede3cc85561f57b550f7e41f589e4531e8a"} Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.960563 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.964931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48jkw" event={"ID":"e1bc708e-493d-4597-9a37-501be2e8641a","Type":"ContainerStarted","Data":"a290f701f683cc1034d01a794078b47d8b2d1ca2e6c9fe64b76b0ec9b751bc1c"} Dec 03 17:17:06 crc kubenswrapper[4758]: I1203 17:17:06.984297 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5czb6" podStartSLOduration=21.361354256 podStartE2EDuration="1m4.984268063s" podCreationTimestamp="2025-12-03 17:16:02 +0000 UTC" firstStartedPulling="2025-12-03 17:16:22.226164919 +0000 UTC m=+1237.427541770" lastFinishedPulling="2025-12-03 17:17:05.849078716 +0000 UTC m=+1281.050455577" observedRunningTime="2025-12-03 17:17:06.97711728 +0000 UTC m=+1282.178494151" watchObservedRunningTime="2025-12-03 17:17:06.984268063 +0000 UTC m=+1282.185644924" Dec 03 17:17:07 crc kubenswrapper[4758]: I1203 17:17:07.005454 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" podStartSLOduration=8.005430036 podStartE2EDuration="8.005430036s" podCreationTimestamp="2025-12-03 17:16:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:06.997208163 +0000 UTC m=+1282.198585044" watchObservedRunningTime="2025-12-03 17:17:07.005430036 +0000 UTC m=+1282.206806897" Dec 03 17:17:07 crc kubenswrapper[4758]: I1203 17:17:07.024317 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.225416886 podStartE2EDuration="1m8.024297746s" podCreationTimestamp="2025-12-03 17:15:59 +0000 UTC" firstStartedPulling="2025-12-03 17:16:01.093325334 +0000 UTC m=+1216.294702195" lastFinishedPulling="2025-12-03 17:17:05.892206194 +0000 UTC m=+1281.093583055" observedRunningTime="2025-12-03 17:17:07.01922532 +0000 UTC m=+1282.220602191" watchObservedRunningTime="2025-12-03 17:17:07.024297746 +0000 UTC m=+1282.225674607" Dec 03 17:17:07 crc kubenswrapper[4758]: I1203 17:17:07.986827 4758 generic.go:334] "Generic (PLEG): container finished" podID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerID="883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74" exitCode=0 Dec 03 17:17:07 crc kubenswrapper[4758]: I1203 17:17:07.987760 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerDied","Data":"883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74"} Dec 03 17:17:08 crc kubenswrapper[4758]: I1203 17:17:08.035642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 17:17:08 crc kubenswrapper[4758]: I1203 17:17:08.515252 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:08 crc kubenswrapper[4758]: E1203 17:17:08.515523 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:17:08 crc kubenswrapper[4758]: E1203 17:17:08.515545 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:17:08 crc kubenswrapper[4758]: E1203 17:17:08.515618 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift podName:946a9b5f-7abf-4b3e-bc51-08a45e73ff30 nodeName:}" failed. No retries permitted until 2025-12-03 17:17:16.515591576 +0000 UTC m=+1291.716968437 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift") pod "swift-storage-0" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30") : configmap "swift-ring-files" not found Dec 03 17:17:08 crc kubenswrapper[4758]: I1203 17:17:08.902952 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bf47b49b7-fjvmb" podUID="91a9316f-a950-4651-98c3-4dd2b0d247bc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 03 17:17:09 crc kubenswrapper[4758]: I1203 17:17:09.002764 4758 generic.go:334] "Generic (PLEG): container finished" podID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerID="ef2a638cba2c7b4383584dce02eb1679141386353eb7c10bd4bd9af144b23c05" exitCode=0 Dec 03 17:17:09 crc kubenswrapper[4758]: I1203 17:17:09.003196 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9c80de0e-2bee-4e3f-8edf-f4a235b01d76","Type":"ContainerDied","Data":"ef2a638cba2c7b4383584dce02eb1679141386353eb7c10bd4bd9af144b23c05"} Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.042460 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48jkw" event={"ID":"e1bc708e-493d-4597-9a37-501be2e8641a","Type":"ContainerStarted","Data":"9804e5f426b099dbf4ce321f0e6e9c4664e9582bdfe2efda45780608e8fea933"} Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.046953 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9c80de0e-2bee-4e3f-8edf-f4a235b01d76","Type":"ContainerStarted","Data":"d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea"} Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.051815 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerStarted","Data":"e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70"} Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.051898 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerStarted","Data":"ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f"} Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.052251 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.071970 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-48jkw" podStartSLOduration=3.425766417 podStartE2EDuration="9.071941557s" podCreationTimestamp="2025-12-03 17:17:04 +0000 UTC" firstStartedPulling="2025-12-03 17:17:06.46896615 +0000 UTC m=+1281.670343011" lastFinishedPulling="2025-12-03 17:17:12.11514129 +0000 UTC m=+1287.316518151" observedRunningTime="2025-12-03 17:17:13.062731977 +0000 UTC m=+1288.264108848" watchObservedRunningTime="2025-12-03 17:17:13.071941557 +0000 UTC m=+1288.273318418" Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.096163 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-5mttb" podStartSLOduration=27.391353198 podStartE2EDuration="1m11.096133822s" podCreationTimestamp="2025-12-03 17:16:02 +0000 UTC" firstStartedPulling="2025-12-03 17:16:22.226851657 +0000 UTC m=+1237.428228518" lastFinishedPulling="2025-12-03 17:17:05.931632281 +0000 UTC m=+1281.133009142" observedRunningTime="2025-12-03 17:17:13.088940117 +0000 UTC m=+1288.290316988" watchObservedRunningTime="2025-12-03 17:17:13.096133822 +0000 UTC m=+1288.297510683" Dec 03 17:17:13 crc kubenswrapper[4758]: I1203 17:17:13.121097 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371957.73371 podStartE2EDuration="1m19.121065897s" podCreationTimestamp="2025-12-03 17:15:54 +0000 UTC" firstStartedPulling="2025-12-03 17:15:56.639938962 +0000 UTC m=+1211.841315823" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:13.114734085 +0000 UTC m=+1288.316110966" watchObservedRunningTime="2025-12-03 17:17:13.121065897 +0000 UTC m=+1288.322442758" Dec 03 17:17:14 crc kubenswrapper[4758]: I1203 17:17:14.060933 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:17:14 crc kubenswrapper[4758]: I1203 17:17:14.866958 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:17:15 crc kubenswrapper[4758]: I1203 17:17:14.961039 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xwrl6"] Dec 03 17:17:15 crc kubenswrapper[4758]: I1203 17:17:14.961471 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-xwrl6" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerName="dnsmasq-dns" containerID="cri-o://0808c7b33148d49a50727b199a344d2d5e5ae74f8fc8755e99598b3aaaffe6b6" gracePeriod=10 Dec 03 17:17:15 crc kubenswrapper[4758]: I1203 17:17:15.819326 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 17:17:15 crc kubenswrapper[4758]: I1203 17:17:15.819497 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.096206 4758 generic.go:334] "Generic (PLEG): container finished" podID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerID="0808c7b33148d49a50727b199a344d2d5e5ae74f8fc8755e99598b3aaaffe6b6" exitCode=0 Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.096754 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xwrl6" event={"ID":"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1","Type":"ContainerDied","Data":"0808c7b33148d49a50727b199a344d2d5e5ae74f8fc8755e99598b3aaaffe6b6"} Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.096831 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xwrl6" event={"ID":"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1","Type":"ContainerDied","Data":"27352ef3d9845af949f946fd24eb9d9c07e6c2b1cdc501333738daff088e67f6"} Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.096866 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27352ef3d9845af949f946fd24eb9d9c07e6c2b1cdc501333738daff088e67f6" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.105755 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.205993 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-dns-svc\") pod \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.206125 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-config\") pod \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.206284 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-sb\") pod \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.206392 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-nb\") pod \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.206550 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsbb8\" (UniqueName: \"kubernetes.io/projected/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-kube-api-access-rsbb8\") pod \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\" (UID: \"a61c673f-5b2a-45b7-96d9-b5cc4c5749b1\") " Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.242659 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-kube-api-access-rsbb8" (OuterVolumeSpecName: "kube-api-access-rsbb8") pod "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" (UID: "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1"). InnerVolumeSpecName "kube-api-access-rsbb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.265704 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-config" (OuterVolumeSpecName: "config") pod "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" (UID: "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.275190 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" (UID: "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.276859 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" (UID: "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.277799 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" (UID: "a61c673f-5b2a-45b7-96d9-b5cc4c5749b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.309757 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.309807 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.309829 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.309844 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsbb8\" (UniqueName: \"kubernetes.io/projected/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-kube-api-access-rsbb8\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.309857 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4758]: I1203 17:17:16.616050 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:16 crc kubenswrapper[4758]: E1203 17:17:16.616638 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:17:16 crc kubenswrapper[4758]: E1203 17:17:16.616785 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:17:16 crc kubenswrapper[4758]: E1203 17:17:16.616855 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift podName:946a9b5f-7abf-4b3e-bc51-08a45e73ff30 nodeName:}" failed. No retries permitted until 2025-12-03 17:17:32.61682457 +0000 UTC m=+1307.818201431 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift") pod "swift-storage-0" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30") : configmap "swift-ring-files" not found Dec 03 17:17:17 crc kubenswrapper[4758]: I1203 17:17:17.106036 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xwrl6" Dec 03 17:17:17 crc kubenswrapper[4758]: I1203 17:17:17.157418 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xwrl6"] Dec 03 17:17:17 crc kubenswrapper[4758]: I1203 17:17:17.164203 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xwrl6"] Dec 03 17:17:19 crc kubenswrapper[4758]: I1203 17:17:19.126420 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" path="/var/lib/kubelet/pods/a61c673f-5b2a-45b7-96d9-b5cc4c5749b1/volumes" Dec 03 17:17:19 crc kubenswrapper[4758]: I1203 17:17:19.665669 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 17:17:21 crc kubenswrapper[4758]: I1203 17:17:21.958647 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.039208 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.541185 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e859-account-create-update-8k259"] Dec 03 17:17:22 crc kubenswrapper[4758]: E1203 17:17:22.542005 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerName="dnsmasq-dns" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.542121 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerName="dnsmasq-dns" Dec 03 17:17:22 crc kubenswrapper[4758]: E1203 17:17:22.542186 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerName="init" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.542247 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerName="init" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.542532 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a61c673f-5b2a-45b7-96d9-b5cc4c5749b1" containerName="dnsmasq-dns" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.543271 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.546414 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.553736 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e859-account-create-update-8k259"] Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.598078 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-c92ds"] Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.600622 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.610837 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c92ds"] Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.667159 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a46406f0-070f-4bb5-890c-50d67da0a338-operator-scripts\") pod \"glance-e859-account-create-update-8k259\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.667269 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fa959d-5438-45b3-a053-a86a205e8da4-operator-scripts\") pod \"glance-db-create-c92ds\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.667318 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2h6x\" (UniqueName: \"kubernetes.io/projected/22fa959d-5438-45b3-a053-a86a205e8da4-kube-api-access-l2h6x\") pod \"glance-db-create-c92ds\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.667641 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b6x6\" (UniqueName: \"kubernetes.io/projected/a46406f0-070f-4bb5-890c-50d67da0a338-kube-api-access-4b6x6\") pod \"glance-e859-account-create-update-8k259\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.769272 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fa959d-5438-45b3-a053-a86a205e8da4-operator-scripts\") pod \"glance-db-create-c92ds\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.769370 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2h6x\" (UniqueName: \"kubernetes.io/projected/22fa959d-5438-45b3-a053-a86a205e8da4-kube-api-access-l2h6x\") pod \"glance-db-create-c92ds\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.769472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b6x6\" (UniqueName: \"kubernetes.io/projected/a46406f0-070f-4bb5-890c-50d67da0a338-kube-api-access-4b6x6\") pod \"glance-e859-account-create-update-8k259\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.769592 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a46406f0-070f-4bb5-890c-50d67da0a338-operator-scripts\") pod \"glance-e859-account-create-update-8k259\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.770333 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a46406f0-070f-4bb5-890c-50d67da0a338-operator-scripts\") pod \"glance-e859-account-create-update-8k259\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.770812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fa959d-5438-45b3-a053-a86a205e8da4-operator-scripts\") pod \"glance-db-create-c92ds\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.795785 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b6x6\" (UniqueName: \"kubernetes.io/projected/a46406f0-070f-4bb5-890c-50d67da0a338-kube-api-access-4b6x6\") pod \"glance-e859-account-create-update-8k259\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.797841 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2h6x\" (UniqueName: \"kubernetes.io/projected/22fa959d-5438-45b3-a053-a86a205e8da4-kube-api-access-l2h6x\") pod \"glance-db-create-c92ds\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " pod="openstack/glance-db-create-c92ds" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.880874 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:22 crc kubenswrapper[4758]: I1203 17:17:22.920892 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c92ds" Dec 03 17:17:23 crc kubenswrapper[4758]: I1203 17:17:23.223269 4758 generic.go:334] "Generic (PLEG): container finished" podID="e1bc708e-493d-4597-9a37-501be2e8641a" containerID="9804e5f426b099dbf4ce321f0e6e9c4664e9582bdfe2efda45780608e8fea933" exitCode=0 Dec 03 17:17:23 crc kubenswrapper[4758]: I1203 17:17:23.223509 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48jkw" event={"ID":"e1bc708e-493d-4597-9a37-501be2e8641a","Type":"ContainerDied","Data":"9804e5f426b099dbf4ce321f0e6e9c4664e9582bdfe2efda45780608e8fea933"} Dec 03 17:17:23 crc kubenswrapper[4758]: I1203 17:17:23.493535 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e859-account-create-update-8k259"] Dec 03 17:17:23 crc kubenswrapper[4758]: W1203 17:17:23.493672 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda46406f0_070f_4bb5_890c_50d67da0a338.slice/crio-1e4d70ea7173d03683d43e01cd1a58877a10c533c133d3905acecab1d525ae7a WatchSource:0}: Error finding container 1e4d70ea7173d03683d43e01cd1a58877a10c533c133d3905acecab1d525ae7a: Status 404 returned error can't find the container with id 1e4d70ea7173d03683d43e01cd1a58877a10c533c133d3905acecab1d525ae7a Dec 03 17:17:23 crc kubenswrapper[4758]: I1203 17:17:23.555102 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c92ds"] Dec 03 17:17:23 crc kubenswrapper[4758]: W1203 17:17:23.572194 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22fa959d_5438_45b3_a053_a86a205e8da4.slice/crio-604b7dd794a61b89c46560e2f05b74842fdfabe08a9c6689c50d2c14ff7b4923 WatchSource:0}: Error finding container 604b7dd794a61b89c46560e2f05b74842fdfabe08a9c6689c50d2c14ff7b4923: Status 404 returned error can't find the container with id 604b7dd794a61b89c46560e2f05b74842fdfabe08a9c6689c50d2c14ff7b4923 Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.237549 4758 generic.go:334] "Generic (PLEG): container finished" podID="22fa959d-5438-45b3-a053-a86a205e8da4" containerID="83f9b29f9b61e08062efbeb4039c8c3a3d37c917f7dfb5c674091aaa5101504a" exitCode=0 Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.237636 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c92ds" event={"ID":"22fa959d-5438-45b3-a053-a86a205e8da4","Type":"ContainerDied","Data":"83f9b29f9b61e08062efbeb4039c8c3a3d37c917f7dfb5c674091aaa5101504a"} Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.237667 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c92ds" event={"ID":"22fa959d-5438-45b3-a053-a86a205e8da4","Type":"ContainerStarted","Data":"604b7dd794a61b89c46560e2f05b74842fdfabe08a9c6689c50d2c14ff7b4923"} Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.242638 4758 generic.go:334] "Generic (PLEG): container finished" podID="a46406f0-070f-4bb5-890c-50d67da0a338" containerID="32e6077c9e48fa700c66888afcec4d226e098bafa50bf0a95011cf9ddf7f603e" exitCode=0 Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.242735 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e859-account-create-update-8k259" event={"ID":"a46406f0-070f-4bb5-890c-50d67da0a338","Type":"ContainerDied","Data":"32e6077c9e48fa700c66888afcec4d226e098bafa50bf0a95011cf9ddf7f603e"} Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.243234 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e859-account-create-update-8k259" event={"ID":"a46406f0-070f-4bb5-890c-50d67da0a338","Type":"ContainerStarted","Data":"1e4d70ea7173d03683d43e01cd1a58877a10c533c133d3905acecab1d525ae7a"} Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.637201 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.713359 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-scripts\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.713428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-swiftconf\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.713534 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4v9t\" (UniqueName: \"kubernetes.io/projected/e1bc708e-493d-4597-9a37-501be2e8641a-kube-api-access-j4v9t\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.713553 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-combined-ca-bundle\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.714819 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1bc708e-493d-4597-9a37-501be2e8641a-etc-swift\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.714855 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-ring-data-devices\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.714948 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-dispersionconf\") pod \"e1bc708e-493d-4597-9a37-501be2e8641a\" (UID: \"e1bc708e-493d-4597-9a37-501be2e8641a\") " Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.715923 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1bc708e-493d-4597-9a37-501be2e8641a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.716020 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.719958 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1bc708e-493d-4597-9a37-501be2e8641a-kube-api-access-j4v9t" (OuterVolumeSpecName: "kube-api-access-j4v9t") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "kube-api-access-j4v9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.722496 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.740016 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-scripts" (OuterVolumeSpecName: "scripts") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.740422 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.743419 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e1bc708e-493d-4597-9a37-501be2e8641a" (UID: "e1bc708e-493d-4597-9a37-501be2e8641a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817342 4758 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1bc708e-493d-4597-9a37-501be2e8641a-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817392 4758 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817403 4758 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817412 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1bc708e-493d-4597-9a37-501be2e8641a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817422 4758 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817431 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4v9t\" (UniqueName: \"kubernetes.io/projected/e1bc708e-493d-4597-9a37-501be2e8641a-kube-api-access-j4v9t\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:24 crc kubenswrapper[4758]: I1203 17:17:24.817442 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bc708e-493d-4597-9a37-501be2e8641a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.256443 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48jkw" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.257231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48jkw" event={"ID":"e1bc708e-493d-4597-9a37-501be2e8641a","Type":"ContainerDied","Data":"a290f701f683cc1034d01a794078b47d8b2d1ca2e6c9fe64b76b0ec9b751bc1c"} Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.257265 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a290f701f683cc1034d01a794078b47d8b2d1ca2e6c9fe64b76b0ec9b751bc1c" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.649379 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c92ds" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.659460 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.736986 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a46406f0-070f-4bb5-890c-50d67da0a338-operator-scripts\") pod \"a46406f0-070f-4bb5-890c-50d67da0a338\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.737451 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a46406f0-070f-4bb5-890c-50d67da0a338-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a46406f0-070f-4bb5-890c-50d67da0a338" (UID: "a46406f0-070f-4bb5-890c-50d67da0a338"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.737717 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fa959d-5438-45b3-a053-a86a205e8da4-operator-scripts\") pod \"22fa959d-5438-45b3-a053-a86a205e8da4\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.737996 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b6x6\" (UniqueName: \"kubernetes.io/projected/a46406f0-070f-4bb5-890c-50d67da0a338-kube-api-access-4b6x6\") pod \"a46406f0-070f-4bb5-890c-50d67da0a338\" (UID: \"a46406f0-070f-4bb5-890c-50d67da0a338\") " Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.738079 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22fa959d-5438-45b3-a053-a86a205e8da4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22fa959d-5438-45b3-a053-a86a205e8da4" (UID: "22fa959d-5438-45b3-a053-a86a205e8da4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.738189 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2h6x\" (UniqueName: \"kubernetes.io/projected/22fa959d-5438-45b3-a053-a86a205e8da4-kube-api-access-l2h6x\") pod \"22fa959d-5438-45b3-a053-a86a205e8da4\" (UID: \"22fa959d-5438-45b3-a053-a86a205e8da4\") " Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.739034 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fa959d-5438-45b3-a053-a86a205e8da4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.739144 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a46406f0-070f-4bb5-890c-50d67da0a338-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.743074 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a46406f0-070f-4bb5-890c-50d67da0a338-kube-api-access-4b6x6" (OuterVolumeSpecName: "kube-api-access-4b6x6") pod "a46406f0-070f-4bb5-890c-50d67da0a338" (UID: "a46406f0-070f-4bb5-890c-50d67da0a338"). InnerVolumeSpecName "kube-api-access-4b6x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.748185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22fa959d-5438-45b3-a053-a86a205e8da4-kube-api-access-l2h6x" (OuterVolumeSpecName: "kube-api-access-l2h6x") pod "22fa959d-5438-45b3-a053-a86a205e8da4" (UID: "22fa959d-5438-45b3-a053-a86a205e8da4"). InnerVolumeSpecName "kube-api-access-l2h6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.840401 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b6x6\" (UniqueName: \"kubernetes.io/projected/a46406f0-070f-4bb5-890c-50d67da0a338-kube-api-access-4b6x6\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:25 crc kubenswrapper[4758]: I1203 17:17:25.840448 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2h6x\" (UniqueName: \"kubernetes.io/projected/22fa959d-5438-45b3-a053-a86a205e8da4-kube-api-access-l2h6x\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.278270 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c92ds" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.278276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c92ds" event={"ID":"22fa959d-5438-45b3-a053-a86a205e8da4","Type":"ContainerDied","Data":"604b7dd794a61b89c46560e2f05b74842fdfabe08a9c6689c50d2c14ff7b4923"} Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.279662 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="604b7dd794a61b89c46560e2f05b74842fdfabe08a9c6689c50d2c14ff7b4923" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.282529 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e859-account-create-update-8k259" event={"ID":"a46406f0-070f-4bb5-890c-50d67da0a338","Type":"ContainerDied","Data":"1e4d70ea7173d03683d43e01cd1a58877a10c533c133d3905acecab1d525ae7a"} Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.282578 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e4d70ea7173d03683d43e01cd1a58877a10c533c133d3905acecab1d525ae7a" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.282617 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e859-account-create-update-8k259" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.769733 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-cpjwp"] Dec 03 17:17:26 crc kubenswrapper[4758]: E1203 17:17:26.770289 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a46406f0-070f-4bb5-890c-50d67da0a338" containerName="mariadb-account-create-update" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.770314 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46406f0-070f-4bb5-890c-50d67da0a338" containerName="mariadb-account-create-update" Dec 03 17:17:26 crc kubenswrapper[4758]: E1203 17:17:26.770338 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1bc708e-493d-4597-9a37-501be2e8641a" containerName="swift-ring-rebalance" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.770347 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1bc708e-493d-4597-9a37-501be2e8641a" containerName="swift-ring-rebalance" Dec 03 17:17:26 crc kubenswrapper[4758]: E1203 17:17:26.770369 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22fa959d-5438-45b3-a053-a86a205e8da4" containerName="mariadb-database-create" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.770377 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="22fa959d-5438-45b3-a053-a86a205e8da4" containerName="mariadb-database-create" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.770703 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="22fa959d-5438-45b3-a053-a86a205e8da4" containerName="mariadb-database-create" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.770731 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a46406f0-070f-4bb5-890c-50d67da0a338" containerName="mariadb-account-create-update" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.770753 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1bc708e-493d-4597-9a37-501be2e8641a" containerName="swift-ring-rebalance" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.771496 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.779739 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-cpjwp"] Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.862693 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84012fbf-daba-42b5-9c41-cd5a5a91da17-operator-scripts\") pod \"keystone-db-create-cpjwp\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.862969 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjtkz\" (UniqueName: \"kubernetes.io/projected/84012fbf-daba-42b5-9c41-cd5a5a91da17-kube-api-access-hjtkz\") pod \"keystone-db-create-cpjwp\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.880052 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8aca-account-create-update-d8h7n"] Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.881712 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.885895 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.889161 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8aca-account-create-update-d8h7n"] Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.964775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84012fbf-daba-42b5-9c41-cd5a5a91da17-operator-scripts\") pod \"keystone-db-create-cpjwp\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.964917 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/b902e850-9e45-4d1d-bc6e-0d71d5830db9-kube-api-access-9kqwm\") pod \"keystone-8aca-account-create-update-d8h7n\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.964973 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjtkz\" (UniqueName: \"kubernetes.io/projected/84012fbf-daba-42b5-9c41-cd5a5a91da17-kube-api-access-hjtkz\") pod \"keystone-db-create-cpjwp\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.965010 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b902e850-9e45-4d1d-bc6e-0d71d5830db9-operator-scripts\") pod \"keystone-8aca-account-create-update-d8h7n\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.971026 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84012fbf-daba-42b5-9c41-cd5a5a91da17-operator-scripts\") pod \"keystone-db-create-cpjwp\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:26 crc kubenswrapper[4758]: I1203 17:17:26.986789 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjtkz\" (UniqueName: \"kubernetes.io/projected/84012fbf-daba-42b5-9c41-cd5a5a91da17-kube-api-access-hjtkz\") pod \"keystone-db-create-cpjwp\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.067295 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/b902e850-9e45-4d1d-bc6e-0d71d5830db9-kube-api-access-9kqwm\") pod \"keystone-8aca-account-create-update-d8h7n\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.067420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b902e850-9e45-4d1d-bc6e-0d71d5830db9-operator-scripts\") pod \"keystone-8aca-account-create-update-d8h7n\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.068892 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b902e850-9e45-4d1d-bc6e-0d71d5830db9-operator-scripts\") pod \"keystone-8aca-account-create-update-d8h7n\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.072546 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-z5cp4"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.073980 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.089487 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/b902e850-9e45-4d1d-bc6e-0d71d5830db9-kube-api-access-9kqwm\") pod \"keystone-8aca-account-create-update-d8h7n\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.090184 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z5cp4"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.091557 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.170545 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/401d7406-8683-47aa-9562-5d3ee99ee30e-operator-scripts\") pod \"placement-db-create-z5cp4\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.171138 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccq68\" (UniqueName: \"kubernetes.io/projected/401d7406-8683-47aa-9562-5d3ee99ee30e-kube-api-access-ccq68\") pod \"placement-db-create-z5cp4\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.202196 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.209748 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-02ea-account-create-update-drbs5"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.217137 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.221935 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.227212 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-02ea-account-create-update-drbs5"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.273318 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/401d7406-8683-47aa-9562-5d3ee99ee30e-operator-scripts\") pod \"placement-db-create-z5cp4\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.273386 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccq68\" (UniqueName: \"kubernetes.io/projected/401d7406-8683-47aa-9562-5d3ee99ee30e-kube-api-access-ccq68\") pod \"placement-db-create-z5cp4\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.280838 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/401d7406-8683-47aa-9562-5d3ee99ee30e-operator-scripts\") pod \"placement-db-create-z5cp4\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.296792 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccq68\" (UniqueName: \"kubernetes.io/projected/401d7406-8683-47aa-9562-5d3ee99ee30e-kube-api-access-ccq68\") pod \"placement-db-create-z5cp4\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.375029 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27b94dc-9078-4c4c-911b-1d1501e71734-operator-scripts\") pod \"placement-02ea-account-create-update-drbs5\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.375481 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tltfg\" (UniqueName: \"kubernetes.io/projected/e27b94dc-9078-4c4c-911b-1d1501e71734-kube-api-access-tltfg\") pod \"placement-02ea-account-create-update-drbs5\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.470470 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.478324 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tltfg\" (UniqueName: \"kubernetes.io/projected/e27b94dc-9078-4c4c-911b-1d1501e71734-kube-api-access-tltfg\") pod \"placement-02ea-account-create-update-drbs5\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.478609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27b94dc-9078-4c4c-911b-1d1501e71734-operator-scripts\") pod \"placement-02ea-account-create-update-drbs5\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.480186 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27b94dc-9078-4c4c-911b-1d1501e71734-operator-scripts\") pod \"placement-02ea-account-create-update-drbs5\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.498341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tltfg\" (UniqueName: \"kubernetes.io/projected/e27b94dc-9078-4c4c-911b-1d1501e71734-kube-api-access-tltfg\") pod \"placement-02ea-account-create-update-drbs5\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.610662 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.656125 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-cpjwp"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.703366 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-qr4sd"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.708148 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.714185 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2c867" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.714409 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 17:17:27 crc kubenswrapper[4758]: W1203 17:17:27.729087 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84012fbf_daba_42b5_9c41_cd5a5a91da17.slice/crio-d78c5323002bbff8c2a4929a5bbba14c45cb4102c8304dbab4d0665e7b41bee8 WatchSource:0}: Error finding container d78c5323002bbff8c2a4929a5bbba14c45cb4102c8304dbab4d0665e7b41bee8: Status 404 returned error can't find the container with id d78c5323002bbff8c2a4929a5bbba14c45cb4102c8304dbab4d0665e7b41bee8 Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.732131 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qr4sd"] Dec 03 17:17:27 crc kubenswrapper[4758]: W1203 17:17:27.778185 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb902e850_9e45_4d1d_bc6e_0d71d5830db9.slice/crio-aa0b98758f3d695d97858c48a2f32868375b1b87f6b9b73ccc36d41e06b376cc WatchSource:0}: Error finding container aa0b98758f3d695d97858c48a2f32868375b1b87f6b9b73ccc36d41e06b376cc: Status 404 returned error can't find the container with id aa0b98758f3d695d97858c48a2f32868375b1b87f6b9b73ccc36d41e06b376cc Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.782045 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8aca-account-create-update-d8h7n"] Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.783555 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-combined-ca-bundle\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.783654 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-config-data\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.783836 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcj97\" (UniqueName: \"kubernetes.io/projected/eb23ba13-deef-46a6-a81d-a223ad56024b-kube-api-access-tcj97\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.783910 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-db-sync-config-data\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.885486 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-config-data\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.885635 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcj97\" (UniqueName: \"kubernetes.io/projected/eb23ba13-deef-46a6-a81d-a223ad56024b-kube-api-access-tcj97\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.885762 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-db-sync-config-data\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.885813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-combined-ca-bundle\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.892087 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-db-sync-config-data\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.892439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-combined-ca-bundle\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.893035 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-config-data\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:27 crc kubenswrapper[4758]: I1203 17:17:27.909105 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcj97\" (UniqueName: \"kubernetes.io/projected/eb23ba13-deef-46a6-a81d-a223ad56024b-kube-api-access-tcj97\") pod \"glance-db-sync-qr4sd\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.043636 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qr4sd" Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.057643 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z5cp4"] Dec 03 17:17:28 crc kubenswrapper[4758]: W1203 17:17:28.068612 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod401d7406_8683_47aa_9562_5d3ee99ee30e.slice/crio-b7fd8c370af1a58aa57032e84202355215df5cff5a90068bb50039876928bcba WatchSource:0}: Error finding container b7fd8c370af1a58aa57032e84202355215df5cff5a90068bb50039876928bcba: Status 404 returned error can't find the container with id b7fd8c370af1a58aa57032e84202355215df5cff5a90068bb50039876928bcba Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.210606 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-02ea-account-create-update-drbs5"] Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.321895 4758 generic.go:334] "Generic (PLEG): container finished" podID="84012fbf-daba-42b5-9c41-cd5a5a91da17" containerID="d634789b8f8c0e90fd890b15f1ea496945238b76512f4197c7fe20fa25e0eaa6" exitCode=0 Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.321960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cpjwp" event={"ID":"84012fbf-daba-42b5-9c41-cd5a5a91da17","Type":"ContainerDied","Data":"d634789b8f8c0e90fd890b15f1ea496945238b76512f4197c7fe20fa25e0eaa6"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.322005 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cpjwp" event={"ID":"84012fbf-daba-42b5-9c41-cd5a5a91da17","Type":"ContainerStarted","Data":"d78c5323002bbff8c2a4929a5bbba14c45cb4102c8304dbab4d0665e7b41bee8"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.323711 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-02ea-account-create-update-drbs5" event={"ID":"e27b94dc-9078-4c4c-911b-1d1501e71734","Type":"ContainerStarted","Data":"bd5b7e9875619287053e2ae058c459c2c361fa463cd469d98b571fe44c8ed547"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.325647 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z5cp4" event={"ID":"401d7406-8683-47aa-9562-5d3ee99ee30e","Type":"ContainerStarted","Data":"b7fd8c370af1a58aa57032e84202355215df5cff5a90068bb50039876928bcba"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.327673 4758 generic.go:334] "Generic (PLEG): container finished" podID="b902e850-9e45-4d1d-bc6e-0d71d5830db9" containerID="23cd376c93f44deb3713feb4e1976b83962a216ee24150ad8c579fa17ac4bff5" exitCode=0 Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.327827 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8aca-account-create-update-d8h7n" event={"ID":"b902e850-9e45-4d1d-bc6e-0d71d5830db9","Type":"ContainerDied","Data":"23cd376c93f44deb3713feb4e1976b83962a216ee24150ad8c579fa17ac4bff5"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.327928 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8aca-account-create-update-d8h7n" event={"ID":"b902e850-9e45-4d1d-bc6e-0d71d5830db9","Type":"ContainerStarted","Data":"aa0b98758f3d695d97858c48a2f32868375b1b87f6b9b73ccc36d41e06b376cc"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.329526 4758 generic.go:334] "Generic (PLEG): container finished" podID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerID="6de2813edf252cbe7e4d319d670dd48259def0afd5b8b085c56a64c56604dd50" exitCode=0 Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.329585 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"070406cc-6d65-4fd0-81d7-264adead6bbd","Type":"ContainerDied","Data":"6de2813edf252cbe7e4d319d670dd48259def0afd5b8b085c56a64c56604dd50"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.337318 4758 generic.go:334] "Generic (PLEG): container finished" podID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerID="085ad6a4f1116fe1014b94a49396f4beca5e6b5a210c06edc52bb5e34006f632" exitCode=0 Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.337376 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e518390e-b4c4-49bb-913f-ddfc331456ab","Type":"ContainerDied","Data":"085ad6a4f1116fe1014b94a49396f4beca5e6b5a210c06edc52bb5e34006f632"} Dec 03 17:17:28 crc kubenswrapper[4758]: I1203 17:17:28.662288 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qr4sd"] Dec 03 17:17:28 crc kubenswrapper[4758]: W1203 17:17:28.673570 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb23ba13_deef_46a6_a81d_a223ad56024b.slice/crio-ba1a227cb327fda2467fe5ec0081b8b44443ef91cf706be0da6dfa7481447ca6 WatchSource:0}: Error finding container ba1a227cb327fda2467fe5ec0081b8b44443ef91cf706be0da6dfa7481447ca6: Status 404 returned error can't find the container with id ba1a227cb327fda2467fe5ec0081b8b44443ef91cf706be0da6dfa7481447ca6 Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.358595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"070406cc-6d65-4fd0-81d7-264adead6bbd","Type":"ContainerStarted","Data":"aa23c937b63cdc7de84a3a65b1ab3cb567e5f8d745fd98201fbefbe82494cbc1"} Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.360366 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.362412 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e518390e-b4c4-49bb-913f-ddfc331456ab","Type":"ContainerStarted","Data":"ffded9f7e8bb729a9659af5430bde139dac85bb5e31acabdfd96c07c75f4d43d"} Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.362645 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.364385 4758 generic.go:334] "Generic (PLEG): container finished" podID="e27b94dc-9078-4c4c-911b-1d1501e71734" containerID="c608fc7bb4625c50cf87800dcbdbdd20deaaadae0ba69d4b513648ae6e5178bd" exitCode=0 Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.364498 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-02ea-account-create-update-drbs5" event={"ID":"e27b94dc-9078-4c4c-911b-1d1501e71734","Type":"ContainerDied","Data":"c608fc7bb4625c50cf87800dcbdbdd20deaaadae0ba69d4b513648ae6e5178bd"} Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.365700 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qr4sd" event={"ID":"eb23ba13-deef-46a6-a81d-a223ad56024b","Type":"ContainerStarted","Data":"ba1a227cb327fda2467fe5ec0081b8b44443ef91cf706be0da6dfa7481447ca6"} Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.367551 4758 generic.go:334] "Generic (PLEG): container finished" podID="401d7406-8683-47aa-9562-5d3ee99ee30e" containerID="e524b6d9bd3a5f4290b198d37c9f8d00e0086aebe49103ce51921e9f8903e0c9" exitCode=0 Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.367644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z5cp4" event={"ID":"401d7406-8683-47aa-9562-5d3ee99ee30e","Type":"ContainerDied","Data":"e524b6d9bd3a5f4290b198d37c9f8d00e0086aebe49103ce51921e9f8903e0c9"} Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.406624 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371939.448193 podStartE2EDuration="1m37.406583226s" podCreationTimestamp="2025-12-03 17:15:52 +0000 UTC" firstStartedPulling="2025-12-03 17:15:54.844890903 +0000 UTC m=+1210.046267764" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:29.405345013 +0000 UTC m=+1304.606721874" watchObservedRunningTime="2025-12-03 17:17:29.406583226 +0000 UTC m=+1304.607960087" Dec 03 17:17:29 crc kubenswrapper[4758]: I1203 17:17:29.545178 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.886245626 podStartE2EDuration="1m36.545153505s" podCreationTimestamp="2025-12-03 17:15:53 +0000 UTC" firstStartedPulling="2025-12-03 17:15:55.962945923 +0000 UTC m=+1211.164322784" lastFinishedPulling="2025-12-03 17:16:53.621853802 +0000 UTC m=+1268.823230663" observedRunningTime="2025-12-03 17:17:29.514101365 +0000 UTC m=+1304.715478226" watchObservedRunningTime="2025-12-03 17:17:29.545153505 +0000 UTC m=+1304.746530366" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.281169 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.377398 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.379215 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-cpjwp" event={"ID":"84012fbf-daba-42b5-9c41-cd5a5a91da17","Type":"ContainerDied","Data":"d78c5323002bbff8c2a4929a5bbba14c45cb4102c8304dbab4d0665e7b41bee8"} Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.379279 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d78c5323002bbff8c2a4929a5bbba14c45cb4102c8304dbab4d0665e7b41bee8" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.381020 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8aca-account-create-update-d8h7n" event={"ID":"b902e850-9e45-4d1d-bc6e-0d71d5830db9","Type":"ContainerDied","Data":"aa0b98758f3d695d97858c48a2f32868375b1b87f6b9b73ccc36d41e06b376cc"} Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.381057 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa0b98758f3d695d97858c48a2f32868375b1b87f6b9b73ccc36d41e06b376cc" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.381418 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8aca-account-create-update-d8h7n" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.451958 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b902e850-9e45-4d1d-bc6e-0d71d5830db9-operator-scripts\") pod \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.452453 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/b902e850-9e45-4d1d-bc6e-0d71d5830db9-kube-api-access-9kqwm\") pod \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\" (UID: \"b902e850-9e45-4d1d-bc6e-0d71d5830db9\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.452944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b902e850-9e45-4d1d-bc6e-0d71d5830db9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b902e850-9e45-4d1d-bc6e-0d71d5830db9" (UID: "b902e850-9e45-4d1d-bc6e-0d71d5830db9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.453617 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b902e850-9e45-4d1d-bc6e-0d71d5830db9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.460021 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b902e850-9e45-4d1d-bc6e-0d71d5830db9-kube-api-access-9kqwm" (OuterVolumeSpecName: "kube-api-access-9kqwm") pod "b902e850-9e45-4d1d-bc6e-0d71d5830db9" (UID: "b902e850-9e45-4d1d-bc6e-0d71d5830db9"). InnerVolumeSpecName "kube-api-access-9kqwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.559594 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84012fbf-daba-42b5-9c41-cd5a5a91da17-operator-scripts\") pod \"84012fbf-daba-42b5-9c41-cd5a5a91da17\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.559770 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjtkz\" (UniqueName: \"kubernetes.io/projected/84012fbf-daba-42b5-9c41-cd5a5a91da17-kube-api-access-hjtkz\") pod \"84012fbf-daba-42b5-9c41-cd5a5a91da17\" (UID: \"84012fbf-daba-42b5-9c41-cd5a5a91da17\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.560299 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84012fbf-daba-42b5-9c41-cd5a5a91da17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84012fbf-daba-42b5-9c41-cd5a5a91da17" (UID: "84012fbf-daba-42b5-9c41-cd5a5a91da17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.561856 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/b902e850-9e45-4d1d-bc6e-0d71d5830db9-kube-api-access-9kqwm\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.561899 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84012fbf-daba-42b5-9c41-cd5a5a91da17-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.565910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84012fbf-daba-42b5-9c41-cd5a5a91da17-kube-api-access-hjtkz" (OuterVolumeSpecName: "kube-api-access-hjtkz") pod "84012fbf-daba-42b5-9c41-cd5a5a91da17" (UID: "84012fbf-daba-42b5-9c41-cd5a5a91da17"). InnerVolumeSpecName "kube-api-access-hjtkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.663383 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjtkz\" (UniqueName: \"kubernetes.io/projected/84012fbf-daba-42b5-9c41-cd5a5a91da17-kube-api-access-hjtkz\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.744041 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.783524 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.873190 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tltfg\" (UniqueName: \"kubernetes.io/projected/e27b94dc-9078-4c4c-911b-1d1501e71734-kube-api-access-tltfg\") pod \"e27b94dc-9078-4c4c-911b-1d1501e71734\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.873346 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27b94dc-9078-4c4c-911b-1d1501e71734-operator-scripts\") pod \"e27b94dc-9078-4c4c-911b-1d1501e71734\" (UID: \"e27b94dc-9078-4c4c-911b-1d1501e71734\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.874085 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27b94dc-9078-4c4c-911b-1d1501e71734-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e27b94dc-9078-4c4c-911b-1d1501e71734" (UID: "e27b94dc-9078-4c4c-911b-1d1501e71734"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.877089 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e27b94dc-9078-4c4c-911b-1d1501e71734-kube-api-access-tltfg" (OuterVolumeSpecName: "kube-api-access-tltfg") pod "e27b94dc-9078-4c4c-911b-1d1501e71734" (UID: "e27b94dc-9078-4c4c-911b-1d1501e71734"). InnerVolumeSpecName "kube-api-access-tltfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.975748 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccq68\" (UniqueName: \"kubernetes.io/projected/401d7406-8683-47aa-9562-5d3ee99ee30e-kube-api-access-ccq68\") pod \"401d7406-8683-47aa-9562-5d3ee99ee30e\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.975923 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/401d7406-8683-47aa-9562-5d3ee99ee30e-operator-scripts\") pod \"401d7406-8683-47aa-9562-5d3ee99ee30e\" (UID: \"401d7406-8683-47aa-9562-5d3ee99ee30e\") " Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.976476 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tltfg\" (UniqueName: \"kubernetes.io/projected/e27b94dc-9078-4c4c-911b-1d1501e71734-kube-api-access-tltfg\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.976505 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27b94dc-9078-4c4c-911b-1d1501e71734-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.976542 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d7406-8683-47aa-9562-5d3ee99ee30e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "401d7406-8683-47aa-9562-5d3ee99ee30e" (UID: "401d7406-8683-47aa-9562-5d3ee99ee30e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:30 crc kubenswrapper[4758]: I1203 17:17:30.979110 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401d7406-8683-47aa-9562-5d3ee99ee30e-kube-api-access-ccq68" (OuterVolumeSpecName: "kube-api-access-ccq68") pod "401d7406-8683-47aa-9562-5d3ee99ee30e" (UID: "401d7406-8683-47aa-9562-5d3ee99ee30e"). InnerVolumeSpecName "kube-api-access-ccq68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.078338 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccq68\" (UniqueName: \"kubernetes.io/projected/401d7406-8683-47aa-9562-5d3ee99ee30e-kube-api-access-ccq68\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.078394 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/401d7406-8683-47aa-9562-5d3ee99ee30e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.394534 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-02ea-account-create-update-drbs5" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.394564 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-02ea-account-create-update-drbs5" event={"ID":"e27b94dc-9078-4c4c-911b-1d1501e71734","Type":"ContainerDied","Data":"bd5b7e9875619287053e2ae058c459c2c361fa463cd469d98b571fe44c8ed547"} Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.394619 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd5b7e9875619287053e2ae058c459c2c361fa463cd469d98b571fe44c8ed547" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.400602 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5cp4" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.401227 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-cpjwp" Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.401501 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z5cp4" event={"ID":"401d7406-8683-47aa-9562-5d3ee99ee30e","Type":"ContainerDied","Data":"b7fd8c370af1a58aa57032e84202355215df5cff5a90068bb50039876928bcba"} Dec 03 17:17:31 crc kubenswrapper[4758]: I1203 17:17:31.401658 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7fd8c370af1a58aa57032e84202355215df5cff5a90068bb50039876928bcba" Dec 03 17:17:32 crc kubenswrapper[4758]: I1203 17:17:32.714558 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:32 crc kubenswrapper[4758]: I1203 17:17:32.735314 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"swift-storage-0\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " pod="openstack/swift-storage-0" Dec 03 17:17:32 crc kubenswrapper[4758]: I1203 17:17:32.827541 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 17:17:33 crc kubenswrapper[4758]: I1203 17:17:33.683352 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:17:33 crc kubenswrapper[4758]: I1203 17:17:33.757635 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"807f57fce17f9d09496cb0a6f17fc8ab10c2dd9c1146877fa4fa1db6de9554ba"} Dec 03 17:17:37 crc kubenswrapper[4758]: I1203 17:17:37.508907 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:17:37 crc kubenswrapper[4758]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 17:17:37 crc kubenswrapper[4758]: > Dec 03 17:17:42 crc kubenswrapper[4758]: I1203 17:17:42.595984 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:17:42 crc kubenswrapper[4758]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 17:17:42 crc kubenswrapper[4758]: > Dec 03 17:17:42 crc kubenswrapper[4758]: I1203 17:17:42.652302 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:17:42 crc kubenswrapper[4758]: I1203 17:17:42.661809 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.023276 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5czb6-config-dp9dg"] Dec 03 17:17:43 crc kubenswrapper[4758]: E1203 17:17:43.023909 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84012fbf-daba-42b5-9c41-cd5a5a91da17" containerName="mariadb-database-create" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.023940 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="84012fbf-daba-42b5-9c41-cd5a5a91da17" containerName="mariadb-database-create" Dec 03 17:17:43 crc kubenswrapper[4758]: E1203 17:17:43.023961 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d7406-8683-47aa-9562-5d3ee99ee30e" containerName="mariadb-database-create" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.023969 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d7406-8683-47aa-9562-5d3ee99ee30e" containerName="mariadb-database-create" Dec 03 17:17:43 crc kubenswrapper[4758]: E1203 17:17:43.023983 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b902e850-9e45-4d1d-bc6e-0d71d5830db9" containerName="mariadb-account-create-update" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.023992 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b902e850-9e45-4d1d-bc6e-0d71d5830db9" containerName="mariadb-account-create-update" Dec 03 17:17:43 crc kubenswrapper[4758]: E1203 17:17:43.024011 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e27b94dc-9078-4c4c-911b-1d1501e71734" containerName="mariadb-account-create-update" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.024017 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e27b94dc-9078-4c4c-911b-1d1501e71734" containerName="mariadb-account-create-update" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.024239 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e27b94dc-9078-4c4c-911b-1d1501e71734" containerName="mariadb-account-create-update" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.024264 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="84012fbf-daba-42b5-9c41-cd5a5a91da17" containerName="mariadb-database-create" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.024275 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b902e850-9e45-4d1d-bc6e-0d71d5830db9" containerName="mariadb-account-create-update" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.024325 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d7406-8683-47aa-9562-5d3ee99ee30e" containerName="mariadb-database-create" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.025253 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.033825 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.035292 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6-config-dp9dg"] Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.197634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-additional-scripts\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.197857 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.197903 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run-ovn\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.197942 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-log-ovn\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.197982 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-scripts\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.198093 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n75k5\" (UniqueName: \"kubernetes.io/projected/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-kube-api-access-n75k5\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.300154 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n75k5\" (UniqueName: \"kubernetes.io/projected/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-kube-api-access-n75k5\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.300251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-additional-scripts\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.300350 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.300391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run-ovn\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.300423 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-log-ovn\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.300472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-scripts\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.301117 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-log-ovn\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.301136 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run-ovn\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.301453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-additional-scripts\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.301766 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.303449 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-scripts\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.325441 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n75k5\" (UniqueName: \"kubernetes.io/projected/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-kube-api-access-n75k5\") pod \"ovn-controller-5czb6-config-dp9dg\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:43 crc kubenswrapper[4758]: I1203 17:17:43.365497 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:44 crc kubenswrapper[4758]: I1203 17:17:44.301980 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.021697 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-njckq"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.023395 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.049945 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e6b7-account-create-update-cxdn6"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.051872 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.058705 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.063418 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-njckq"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.067762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d868cfc0-6d71-4170-b50d-5f6a411111c8-operator-scripts\") pod \"barbican-e6b7-account-create-update-cxdn6\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.067867 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8fgd\" (UniqueName: \"kubernetes.io/projected/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-kube-api-access-h8fgd\") pod \"cinder-db-create-njckq\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.068149 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-operator-scripts\") pod \"cinder-db-create-njckq\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.068200 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nwfb\" (UniqueName: \"kubernetes.io/projected/d868cfc0-6d71-4170-b50d-5f6a411111c8-kube-api-access-5nwfb\") pod \"barbican-e6b7-account-create-update-cxdn6\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.194555 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8fgd\" (UniqueName: \"kubernetes.io/projected/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-kube-api-access-h8fgd\") pod \"cinder-db-create-njckq\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.194604 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-operator-scripts\") pod \"cinder-db-create-njckq\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.198849 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-operator-scripts\") pod \"cinder-db-create-njckq\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.198893 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nwfb\" (UniqueName: \"kubernetes.io/projected/d868cfc0-6d71-4170-b50d-5f6a411111c8-kube-api-access-5nwfb\") pod \"barbican-e6b7-account-create-update-cxdn6\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.199120 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d868cfc0-6d71-4170-b50d-5f6a411111c8-operator-scripts\") pod \"barbican-e6b7-account-create-update-cxdn6\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.199377 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e6b7-account-create-update-cxdn6"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.220968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d868cfc0-6d71-4170-b50d-5f6a411111c8-operator-scripts\") pod \"barbican-e6b7-account-create-update-cxdn6\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.240738 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nwfb\" (UniqueName: \"kubernetes.io/projected/d868cfc0-6d71-4170-b50d-5f6a411111c8-kube-api-access-5nwfb\") pod \"barbican-e6b7-account-create-update-cxdn6\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.237015 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-14bb-account-create-update-s96dt"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.267085 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8fgd\" (UniqueName: \"kubernetes.io/projected/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-kube-api-access-h8fgd\") pod \"cinder-db-create-njckq\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.275173 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qrldg"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.276071 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-14bb-account-create-update-s96dt"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.276102 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qrldg"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.276239 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.276905 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.311455 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.322121 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20469a6a-4abc-4e0b-886f-ba199ee5caa3-operator-scripts\") pod \"barbican-db-create-qrldg\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.322175 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/428ee377-ea6d-4897-890a-6676a318c95a-operator-scripts\") pod \"cinder-14bb-account-create-update-s96dt\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.322200 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dxnz\" (UniqueName: \"kubernetes.io/projected/428ee377-ea6d-4897-890a-6676a318c95a-kube-api-access-5dxnz\") pod \"cinder-14bb-account-create-update-s96dt\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.322235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nc4n\" (UniqueName: \"kubernetes.io/projected/20469a6a-4abc-4e0b-886f-ba199ee5caa3-kube-api-access-7nc4n\") pod \"barbican-db-create-qrldg\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.372419 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-njckq" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.394914 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.412358 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.421182 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-stqc7"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.423040 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.424668 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20469a6a-4abc-4e0b-886f-ba199ee5caa3-operator-scripts\") pod \"barbican-db-create-qrldg\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.424762 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/428ee377-ea6d-4897-890a-6676a318c95a-operator-scripts\") pod \"cinder-14bb-account-create-update-s96dt\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.424806 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dxnz\" (UniqueName: \"kubernetes.io/projected/428ee377-ea6d-4897-890a-6676a318c95a-kube-api-access-5dxnz\") pod \"cinder-14bb-account-create-update-s96dt\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.424843 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nc4n\" (UniqueName: \"kubernetes.io/projected/20469a6a-4abc-4e0b-886f-ba199ee5caa3-kube-api-access-7nc4n\") pod \"barbican-db-create-qrldg\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.426496 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20469a6a-4abc-4e0b-886f-ba199ee5caa3-operator-scripts\") pod \"barbican-db-create-qrldg\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.438534 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.439122 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.439894 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9h99t" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.440034 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.450327 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/428ee377-ea6d-4897-890a-6676a318c95a-operator-scripts\") pod \"cinder-14bb-account-create-update-s96dt\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.452980 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-stqc7"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.465985 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dxnz\" (UniqueName: \"kubernetes.io/projected/428ee377-ea6d-4897-890a-6676a318c95a-kube-api-access-5dxnz\") pod \"cinder-14bb-account-create-update-s96dt\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.495470 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nc4n\" (UniqueName: \"kubernetes.io/projected/20469a6a-4abc-4e0b-886f-ba199ee5caa3-kube-api-access-7nc4n\") pod \"barbican-db-create-qrldg\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.562552 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2kcmv"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.566594 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.575589 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2kcmv"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.633502 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-config-data\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.633622 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j946t\" (UniqueName: \"kubernetes.io/projected/8e2e52f9-0802-4065-9630-2a2200060db4-kube-api-access-j946t\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.633702 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-combined-ca-bundle\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.664249 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qrldg" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.684138 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-988a-account-create-update-6fjck"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.685826 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.691408 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.692207 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.700114 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-988a-account-create-update-6fjck"] Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.735562 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-operator-scripts\") pod \"neutron-db-create-2kcmv\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.735735 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbk6k\" (UniqueName: \"kubernetes.io/projected/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-kube-api-access-qbk6k\") pod \"neutron-db-create-2kcmv\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.735867 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-config-data\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.735950 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j946t\" (UniqueName: \"kubernetes.io/projected/8e2e52f9-0802-4065-9630-2a2200060db4-kube-api-access-j946t\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.736545 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-combined-ca-bundle\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.755256 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-combined-ca-bundle\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.768303 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-config-data\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.779733 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j946t\" (UniqueName: \"kubernetes.io/projected/8e2e52f9-0802-4065-9630-2a2200060db4-kube-api-access-j946t\") pod \"keystone-db-sync-stqc7\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.841298 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbk6k\" (UniqueName: \"kubernetes.io/projected/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-kube-api-access-qbk6k\") pod \"neutron-db-create-2kcmv\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.841400 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-operator-scripts\") pod \"neutron-988a-account-create-update-6fjck\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.841442 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbt5\" (UniqueName: \"kubernetes.io/projected/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-kube-api-access-sgbt5\") pod \"neutron-988a-account-create-update-6fjck\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.841582 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-operator-scripts\") pod \"neutron-db-create-2kcmv\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.842534 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-operator-scripts\") pod \"neutron-db-create-2kcmv\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.866517 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbk6k\" (UniqueName: \"kubernetes.io/projected/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-kube-api-access-qbk6k\") pod \"neutron-db-create-2kcmv\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.887180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2kcmv" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.946816 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-operator-scripts\") pod \"neutron-988a-account-create-update-6fjck\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.946902 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbt5\" (UniqueName: \"kubernetes.io/projected/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-kube-api-access-sgbt5\") pod \"neutron-988a-account-create-update-6fjck\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.947772 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-operator-scripts\") pod \"neutron-988a-account-create-update-6fjck\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:45 crc kubenswrapper[4758]: I1203 17:17:45.969803 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbt5\" (UniqueName: \"kubernetes.io/projected/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-kube-api-access-sgbt5\") pod \"neutron-988a-account-create-update-6fjck\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:46 crc kubenswrapper[4758]: I1203 17:17:46.013827 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:17:46 crc kubenswrapper[4758]: I1203 17:17:46.060440 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stqc7" Dec 03 17:17:47 crc kubenswrapper[4758]: I1203 17:17:47.525496 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:17:47 crc kubenswrapper[4758]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 17:17:47 crc kubenswrapper[4758]: > Dec 03 17:17:52 crc kubenswrapper[4758]: E1203 17:17:52.407514 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 03 17:17:52 crc kubenswrapper[4758]: E1203 17:17:52.410374 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tcj97,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-qr4sd_openstack(eb23ba13-deef-46a6-a81d-a223ad56024b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:17:52 crc kubenswrapper[4758]: E1203 17:17:52.411754 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-qr4sd" podUID="eb23ba13-deef-46a6-a81d-a223ad56024b" Dec 03 17:17:52 crc kubenswrapper[4758]: I1203 17:17:52.488673 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:17:52 crc kubenswrapper[4758]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 17:17:52 crc kubenswrapper[4758]: > Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.068588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19"} Dec 03 17:17:53 crc kubenswrapper[4758]: E1203 17:17:53.081729 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-qr4sd" podUID="eb23ba13-deef-46a6-a81d-a223ad56024b" Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.805758 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-njckq"] Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.824431 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e6b7-account-create-update-cxdn6"] Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.832936 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-988a-account-create-update-6fjck"] Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.844858 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2kcmv"] Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.855760 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6-config-dp9dg"] Dec 03 17:17:53 crc kubenswrapper[4758]: W1203 17:17:53.859661 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod428ee377_ea6d_4897_890a_6676a318c95a.slice/crio-63b00bee2eaf0f6558d377b5f684870d48e1d780311b068d69c9f62f2403da66 WatchSource:0}: Error finding container 63b00bee2eaf0f6558d377b5f684870d48e1d780311b068d69c9f62f2403da66: Status 404 returned error can't find the container with id 63b00bee2eaf0f6558d377b5f684870d48e1d780311b068d69c9f62f2403da66 Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.876036 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-14bb-account-create-update-s96dt"] Dec 03 17:17:53 crc kubenswrapper[4758]: I1203 17:17:53.903358 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qrldg"] Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.013617 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-stqc7"] Dec 03 17:17:54 crc kubenswrapper[4758]: W1203 17:17:54.055829 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e2e52f9_0802_4065_9630_2a2200060db4.slice/crio-7914c3f813ccfdc70f5f97d8ec44ee6e028ad454a9df801183820f2772960f30 WatchSource:0}: Error finding container 7914c3f813ccfdc70f5f97d8ec44ee6e028ad454a9df801183820f2772960f30: Status 404 returned error can't find the container with id 7914c3f813ccfdc70f5f97d8ec44ee6e028ad454a9df801183820f2772960f30 Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.087855 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qrldg" event={"ID":"20469a6a-4abc-4e0b-886f-ba199ee5caa3","Type":"ContainerStarted","Data":"068ded0fb5af62df7ccfb5ce2e1dc352e359ba5b78d376d451fdcbf337fe5bd0"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.096264 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-njckq" event={"ID":"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7","Type":"ContainerStarted","Data":"310bda9199d4a77b2c39ba206464ebe6c3f20939f497028b1956cbc964da1717"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.108607 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stqc7" event={"ID":"8e2e52f9-0802-4065-9630-2a2200060db4","Type":"ContainerStarted","Data":"7914c3f813ccfdc70f5f97d8ec44ee6e028ad454a9df801183820f2772960f30"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.116554 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e6b7-account-create-update-cxdn6" event={"ID":"d868cfc0-6d71-4170-b50d-5f6a411111c8","Type":"ContainerStarted","Data":"5736bff2379d89c300808759990ad1696836750a20cc0dec570e57b102ef0523"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.124490 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-dp9dg" event={"ID":"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1","Type":"ContainerStarted","Data":"aca490b27ad60c4f807c60fcf599105ab7c9a90b905b81b1d54f627489b919d4"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.130244 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-14bb-account-create-update-s96dt" event={"ID":"428ee377-ea6d-4897-890a-6676a318c95a","Type":"ContainerStarted","Data":"63b00bee2eaf0f6558d377b5f684870d48e1d780311b068d69c9f62f2403da66"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.135182 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.135233 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.135243 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.140231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2kcmv" event={"ID":"0a0e71a2-24ae-4842-935e-5ed7a9bc7145","Type":"ContainerStarted","Data":"0cd8f31014fd1c4de1be926bfcf034428758d73a72501263d2d7e6074a4defd8"} Dec 03 17:17:54 crc kubenswrapper[4758]: I1203 17:17:54.144975 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988a-account-create-update-6fjck" event={"ID":"fde0a172-de16-4feb-a2e0-c51ccb8f41e2","Type":"ContainerStarted","Data":"cf6dd04855a4edf58650daaf020b33e5d4a1aea6fcf50a44bb9ddd26147ede3f"} Dec 03 17:17:55 crc kubenswrapper[4758]: E1203 17:17:55.084706 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea7d4cc9_0b16_429d_a3b0_9a471c98a8f7.slice/crio-067f55e5df424ba5ecc4abf589ef0908d61bb0bb9e7faae3ba23b2e137ee2b4f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.167381 4758 generic.go:334] "Generic (PLEG): container finished" podID="ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" containerID="067f55e5df424ba5ecc4abf589ef0908d61bb0bb9e7faae3ba23b2e137ee2b4f" exitCode=0 Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.167546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-njckq" event={"ID":"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7","Type":"ContainerDied","Data":"067f55e5df424ba5ecc4abf589ef0908d61bb0bb9e7faae3ba23b2e137ee2b4f"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.173368 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988a-account-create-update-6fjck" event={"ID":"fde0a172-de16-4feb-a2e0-c51ccb8f41e2","Type":"ContainerStarted","Data":"e1b3122e4ed5eb97bfcaa944671ba839285f8b7c938d8f2fddbfabb7cdc55868"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.182867 4758 generic.go:334] "Generic (PLEG): container finished" podID="d868cfc0-6d71-4170-b50d-5f6a411111c8" containerID="412987fdb9dbd3120fe67645f65ca5836f91df6e24f7be18c9b5b46a4e6cccb1" exitCode=0 Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.183004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e6b7-account-create-update-cxdn6" event={"ID":"d868cfc0-6d71-4170-b50d-5f6a411111c8","Type":"ContainerDied","Data":"412987fdb9dbd3120fe67645f65ca5836f91df6e24f7be18c9b5b46a4e6cccb1"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.187045 4758 generic.go:334] "Generic (PLEG): container finished" podID="6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" containerID="dc93f50cb6ec038cc05b073430b2e0b33bcdb3500813955c37b1871f39af4d2d" exitCode=0 Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.187148 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-dp9dg" event={"ID":"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1","Type":"ContainerDied","Data":"dc93f50cb6ec038cc05b073430b2e0b33bcdb3500813955c37b1871f39af4d2d"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.196730 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-14bb-account-create-update-s96dt" event={"ID":"428ee377-ea6d-4897-890a-6676a318c95a","Type":"ContainerStarted","Data":"43db26e3f86de23e8be92405726fa2e1dfd092fdd6eed2ce63132cd949d1945d"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.201812 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2kcmv" event={"ID":"0a0e71a2-24ae-4842-935e-5ed7a9bc7145","Type":"ContainerStarted","Data":"a74f15a6d511a8bec6e690deb81da4dadb59f18bcc62fea8232cd90e22e04883"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.205961 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qrldg" event={"ID":"20469a6a-4abc-4e0b-886f-ba199ee5caa3","Type":"ContainerStarted","Data":"1f7ba867db015c485ed3d0570727a2b99fbe07212ced2bb21d6f5f4e010efe7e"} Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.355005 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-qrldg" podStartSLOduration=10.354961439 podStartE2EDuration="10.354961439s" podCreationTimestamp="2025-12-03 17:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:55.328136063 +0000 UTC m=+1330.529512924" watchObservedRunningTime="2025-12-03 17:17:55.354961439 +0000 UTC m=+1330.556338300" Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.389878 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-2kcmv" podStartSLOduration=10.389848842 podStartE2EDuration="10.389848842s" podCreationTimestamp="2025-12-03 17:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:55.367016584 +0000 UTC m=+1330.568393445" watchObservedRunningTime="2025-12-03 17:17:55.389848842 +0000 UTC m=+1330.591225703" Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.413918 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-14bb-account-create-update-s96dt" podStartSLOduration=10.413889452 podStartE2EDuration="10.413889452s" podCreationTimestamp="2025-12-03 17:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:55.407317124 +0000 UTC m=+1330.608693995" watchObservedRunningTime="2025-12-03 17:17:55.413889452 +0000 UTC m=+1330.615266313" Dec 03 17:17:55 crc kubenswrapper[4758]: I1203 17:17:55.434166 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-988a-account-create-update-6fjck" podStartSLOduration=10.43413277 podStartE2EDuration="10.43413277s" podCreationTimestamp="2025-12-03 17:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:55.421417726 +0000 UTC m=+1330.622794587" watchObservedRunningTime="2025-12-03 17:17:55.43413277 +0000 UTC m=+1330.635509631" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.231695 4758 generic.go:334] "Generic (PLEG): container finished" podID="20469a6a-4abc-4e0b-886f-ba199ee5caa3" containerID="1f7ba867db015c485ed3d0570727a2b99fbe07212ced2bb21d6f5f4e010efe7e" exitCode=0 Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.232631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qrldg" event={"ID":"20469a6a-4abc-4e0b-886f-ba199ee5caa3","Type":"ContainerDied","Data":"1f7ba867db015c485ed3d0570727a2b99fbe07212ced2bb21d6f5f4e010efe7e"} Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.238481 4758 generic.go:334] "Generic (PLEG): container finished" podID="0a0e71a2-24ae-4842-935e-5ed7a9bc7145" containerID="a74f15a6d511a8bec6e690deb81da4dadb59f18bcc62fea8232cd90e22e04883" exitCode=0 Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.239748 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2kcmv" event={"ID":"0a0e71a2-24ae-4842-935e-5ed7a9bc7145","Type":"ContainerDied","Data":"a74f15a6d511a8bec6e690deb81da4dadb59f18bcc62fea8232cd90e22e04883"} Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.746196 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-njckq" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.757986 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.769316 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.806913 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d868cfc0-6d71-4170-b50d-5f6a411111c8-operator-scripts\") pod \"d868cfc0-6d71-4170-b50d-5f6a411111c8\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807050 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8fgd\" (UniqueName: \"kubernetes.io/projected/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-kube-api-access-h8fgd\") pod \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807092 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-operator-scripts\") pod \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\" (UID: \"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807139 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-additional-scripts\") pod \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807214 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-scripts\") pod \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nwfb\" (UniqueName: \"kubernetes.io/projected/d868cfc0-6d71-4170-b50d-5f6a411111c8-kube-api-access-5nwfb\") pod \"d868cfc0-6d71-4170-b50d-5f6a411111c8\" (UID: \"d868cfc0-6d71-4170-b50d-5f6a411111c8\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807307 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n75k5\" (UniqueName: \"kubernetes.io/projected/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-kube-api-access-n75k5\") pod \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807342 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-log-ovn\") pod \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807363 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run-ovn\") pod \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807429 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run\") pod \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\" (UID: \"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1\") " Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.807906 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run" (OuterVolumeSpecName: "var-run") pod "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" (UID: "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.808698 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d868cfc0-6d71-4170-b50d-5f6a411111c8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d868cfc0-6d71-4170-b50d-5f6a411111c8" (UID: "d868cfc0-6d71-4170-b50d-5f6a411111c8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.809241 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" (UID: "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.809722 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" (UID: "ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.809548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" (UID: "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.810424 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" (UID: "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.810984 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-scripts" (OuterVolumeSpecName: "scripts") pod "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" (UID: "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.824025 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d868cfc0-6d71-4170-b50d-5f6a411111c8-kube-api-access-5nwfb" (OuterVolumeSpecName: "kube-api-access-5nwfb") pod "d868cfc0-6d71-4170-b50d-5f6a411111c8" (UID: "d868cfc0-6d71-4170-b50d-5f6a411111c8"). InnerVolumeSpecName "kube-api-access-5nwfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.824255 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-kube-api-access-n75k5" (OuterVolumeSpecName: "kube-api-access-n75k5") pod "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" (UID: "6d2a39e2-b7a8-45c2-8542-e01b4c5677c1"). InnerVolumeSpecName "kube-api-access-n75k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.827066 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-kube-api-access-h8fgd" (OuterVolumeSpecName: "kube-api-access-h8fgd") pod "ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" (UID: "ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7"). InnerVolumeSpecName "kube-api-access-h8fgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910195 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910255 4758 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910275 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910288 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nwfb\" (UniqueName: \"kubernetes.io/projected/d868cfc0-6d71-4170-b50d-5f6a411111c8-kube-api-access-5nwfb\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910305 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n75k5\" (UniqueName: \"kubernetes.io/projected/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-kube-api-access-n75k5\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910318 4758 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910333 4758 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910345 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910356 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d868cfc0-6d71-4170-b50d-5f6a411111c8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:56 crc kubenswrapper[4758]: I1203 17:17:56.910367 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8fgd\" (UniqueName: \"kubernetes.io/projected/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7-kube-api-access-h8fgd\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.255180 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025"} Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.259187 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-njckq" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.259207 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-njckq" event={"ID":"ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7","Type":"ContainerDied","Data":"310bda9199d4a77b2c39ba206464ebe6c3f20939f497028b1956cbc964da1717"} Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.259282 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="310bda9199d4a77b2c39ba206464ebe6c3f20939f497028b1956cbc964da1717" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.264109 4758 generic.go:334] "Generic (PLEG): container finished" podID="fde0a172-de16-4feb-a2e0-c51ccb8f41e2" containerID="e1b3122e4ed5eb97bfcaa944671ba839285f8b7c938d8f2fddbfabb7cdc55868" exitCode=0 Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.264217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988a-account-create-update-6fjck" event={"ID":"fde0a172-de16-4feb-a2e0-c51ccb8f41e2","Type":"ContainerDied","Data":"e1b3122e4ed5eb97bfcaa944671ba839285f8b7c938d8f2fddbfabb7cdc55868"} Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.268065 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6b7-account-create-update-cxdn6" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.268632 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e6b7-account-create-update-cxdn6" event={"ID":"d868cfc0-6d71-4170-b50d-5f6a411111c8","Type":"ContainerDied","Data":"5736bff2379d89c300808759990ad1696836750a20cc0dec570e57b102ef0523"} Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.268662 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5736bff2379d89c300808759990ad1696836750a20cc0dec570e57b102ef0523" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.272097 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-dp9dg" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.272125 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-dp9dg" event={"ID":"6d2a39e2-b7a8-45c2-8542-e01b4c5677c1","Type":"ContainerDied","Data":"aca490b27ad60c4f807c60fcf599105ab7c9a90b905b81b1d54f627489b919d4"} Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.272176 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aca490b27ad60c4f807c60fcf599105ab7c9a90b905b81b1d54f627489b919d4" Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.276259 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-14bb-account-create-update-s96dt" event={"ID":"428ee377-ea6d-4897-890a-6676a318c95a","Type":"ContainerDied","Data":"43db26e3f86de23e8be92405726fa2e1dfd092fdd6eed2ce63132cd949d1945d"} Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.276190 4758 generic.go:334] "Generic (PLEG): container finished" podID="428ee377-ea6d-4897-890a-6676a318c95a" containerID="43db26e3f86de23e8be92405726fa2e1dfd092fdd6eed2ce63132cd949d1945d" exitCode=0 Dec 03 17:17:57 crc kubenswrapper[4758]: I1203 17:17:57.520343 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-5czb6" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.012925 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5czb6-config-dp9dg"] Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.023409 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5czb6-config-dp9dg"] Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.306034 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5czb6-config-2hn4j"] Dec 03 17:17:58 crc kubenswrapper[4758]: E1203 17:17:58.306591 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" containerName="mariadb-database-create" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.306613 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" containerName="mariadb-database-create" Dec 03 17:17:58 crc kubenswrapper[4758]: E1203 17:17:58.306625 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" containerName="ovn-config" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.306633 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" containerName="ovn-config" Dec 03 17:17:58 crc kubenswrapper[4758]: E1203 17:17:58.306674 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d868cfc0-6d71-4170-b50d-5f6a411111c8" containerName="mariadb-account-create-update" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.306705 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d868cfc0-6d71-4170-b50d-5f6a411111c8" containerName="mariadb-account-create-update" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.306976 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d868cfc0-6d71-4170-b50d-5f6a411111c8" containerName="mariadb-account-create-update" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.306993 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" containerName="ovn-config" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.307025 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" containerName="mariadb-database-create" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.307919 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.313658 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.337916 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6-config-2hn4j"] Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.388395 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-scripts\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.388543 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.388603 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-log-ovn\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.388654 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-additional-scripts\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.389034 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run-ovn\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.389134 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frl42\" (UniqueName: \"kubernetes.io/projected/9ed875bd-af54-48cb-91ca-b6f92fda869f-kube-api-access-frl42\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494062 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run-ovn\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frl42\" (UniqueName: \"kubernetes.io/projected/9ed875bd-af54-48cb-91ca-b6f92fda869f-kube-api-access-frl42\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494249 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-scripts\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494429 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494523 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-log-ovn\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494608 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-additional-scripts\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494662 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run-ovn\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494891 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-log-ovn\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.494926 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.496628 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-additional-scripts\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.497362 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-scripts\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.530748 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frl42\" (UniqueName: \"kubernetes.io/projected/9ed875bd-af54-48cb-91ca-b6f92fda869f-kube-api-access-frl42\") pod \"ovn-controller-5czb6-config-2hn4j\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:58 crc kubenswrapper[4758]: I1203 17:17:58.644921 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:17:59 crc kubenswrapper[4758]: I1203 17:17:59.131640 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d2a39e2-b7a8-45c2-8542-e01b4c5677c1" path="/var/lib/kubelet/pods/6d2a39e2-b7a8-45c2-8542-e01b4c5677c1/volumes" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.572260 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.646243 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dxnz\" (UniqueName: \"kubernetes.io/projected/428ee377-ea6d-4897-890a-6676a318c95a-kube-api-access-5dxnz\") pod \"428ee377-ea6d-4897-890a-6676a318c95a\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.646431 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/428ee377-ea6d-4897-890a-6676a318c95a-operator-scripts\") pod \"428ee377-ea6d-4897-890a-6676a318c95a\" (UID: \"428ee377-ea6d-4897-890a-6676a318c95a\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.647525 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/428ee377-ea6d-4897-890a-6676a318c95a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "428ee377-ea6d-4897-890a-6676a318c95a" (UID: "428ee377-ea6d-4897-890a-6676a318c95a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.655468 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/428ee377-ea6d-4897-890a-6676a318c95a-kube-api-access-5dxnz" (OuterVolumeSpecName: "kube-api-access-5dxnz") pod "428ee377-ea6d-4897-890a-6676a318c95a" (UID: "428ee377-ea6d-4897-890a-6676a318c95a"). InnerVolumeSpecName "kube-api-access-5dxnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.749376 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/428ee377-ea6d-4897-890a-6676a318c95a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.749458 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dxnz\" (UniqueName: \"kubernetes.io/projected/428ee377-ea6d-4897-890a-6676a318c95a-kube-api-access-5dxnz\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.791132 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.840955 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qrldg" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.850265 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20469a6a-4abc-4e0b-886f-ba199ee5caa3-operator-scripts\") pod \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.850379 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nc4n\" (UniqueName: \"kubernetes.io/projected/20469a6a-4abc-4e0b-886f-ba199ee5caa3-kube-api-access-7nc4n\") pod \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\" (UID: \"20469a6a-4abc-4e0b-886f-ba199ee5caa3\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.850414 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2kcmv" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.850448 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbt5\" (UniqueName: \"kubernetes.io/projected/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-kube-api-access-sgbt5\") pod \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.850541 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-operator-scripts\") pod \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\" (UID: \"fde0a172-de16-4feb-a2e0-c51ccb8f41e2\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.851821 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fde0a172-de16-4feb-a2e0-c51ccb8f41e2" (UID: "fde0a172-de16-4feb-a2e0-c51ccb8f41e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.852463 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20469a6a-4abc-4e0b-886f-ba199ee5caa3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20469a6a-4abc-4e0b-886f-ba199ee5caa3" (UID: "20469a6a-4abc-4e0b-886f-ba199ee5caa3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.865176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-kube-api-access-sgbt5" (OuterVolumeSpecName: "kube-api-access-sgbt5") pod "fde0a172-de16-4feb-a2e0-c51ccb8f41e2" (UID: "fde0a172-de16-4feb-a2e0-c51ccb8f41e2"). InnerVolumeSpecName "kube-api-access-sgbt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.865880 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20469a6a-4abc-4e0b-886f-ba199ee5caa3-kube-api-access-7nc4n" (OuterVolumeSpecName: "kube-api-access-7nc4n") pod "20469a6a-4abc-4e0b-886f-ba199ee5caa3" (UID: "20469a6a-4abc-4e0b-886f-ba199ee5caa3"). InnerVolumeSpecName "kube-api-access-7nc4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.939370 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6-config-2hn4j"] Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.954841 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbk6k\" (UniqueName: \"kubernetes.io/projected/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-kube-api-access-qbk6k\") pod \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.955054 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-operator-scripts\") pod \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\" (UID: \"0a0e71a2-24ae-4842-935e-5ed7a9bc7145\") " Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.955594 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.955619 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20469a6a-4abc-4e0b-886f-ba199ee5caa3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.955629 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nc4n\" (UniqueName: \"kubernetes.io/projected/20469a6a-4abc-4e0b-886f-ba199ee5caa3-kube-api-access-7nc4n\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.955640 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbt5\" (UniqueName: \"kubernetes.io/projected/fde0a172-de16-4feb-a2e0-c51ccb8f41e2-kube-api-access-sgbt5\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.956206 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a0e71a2-24ae-4842-935e-5ed7a9bc7145" (UID: "0a0e71a2-24ae-4842-935e-5ed7a9bc7145"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:05 crc kubenswrapper[4758]: I1203 17:18:05.962598 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-kube-api-access-qbk6k" (OuterVolumeSpecName: "kube-api-access-qbk6k") pod "0a0e71a2-24ae-4842-935e-5ed7a9bc7145" (UID: "0a0e71a2-24ae-4842-935e-5ed7a9bc7145"). InnerVolumeSpecName "kube-api-access-qbk6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.057536 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbk6k\" (UniqueName: \"kubernetes.io/projected/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-kube-api-access-qbk6k\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.057574 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a0e71a2-24ae-4842-935e-5ed7a9bc7145-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.422342 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.422856 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.422876 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.438225 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2kcmv" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.438587 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2kcmv" event={"ID":"0a0e71a2-24ae-4842-935e-5ed7a9bc7145","Type":"ContainerDied","Data":"0cd8f31014fd1c4de1be926bfcf034428758d73a72501263d2d7e6074a4defd8"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.438669 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cd8f31014fd1c4de1be926bfcf034428758d73a72501263d2d7e6074a4defd8" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.446494 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qrldg" event={"ID":"20469a6a-4abc-4e0b-886f-ba199ee5caa3","Type":"ContainerDied","Data":"068ded0fb5af62df7ccfb5ce2e1dc352e359ba5b78d376d451fdcbf337fe5bd0"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.446555 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="068ded0fb5af62df7ccfb5ce2e1dc352e359ba5b78d376d451fdcbf337fe5bd0" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.446664 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qrldg" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.464286 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988a-account-create-update-6fjck" event={"ID":"fde0a172-de16-4feb-a2e0-c51ccb8f41e2","Type":"ContainerDied","Data":"cf6dd04855a4edf58650daaf020b33e5d4a1aea6fcf50a44bb9ddd26147ede3f"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.465331 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf6dd04855a4edf58650daaf020b33e5d4a1aea6fcf50a44bb9ddd26147ede3f" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.465973 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-988a-account-create-update-6fjck" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.469258 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-2hn4j" event={"ID":"9ed875bd-af54-48cb-91ca-b6f92fda869f","Type":"ContainerStarted","Data":"6d70a91ca70352bcd228584cfc0a35eeeed5d96455696994c096ca68be4c414f"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.469314 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-2hn4j" event={"ID":"9ed875bd-af54-48cb-91ca-b6f92fda869f","Type":"ContainerStarted","Data":"8800a232afab58d607dcc7069076495a9e2397accbf7ddeb59dd8175f8f710d5"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.471631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stqc7" event={"ID":"8e2e52f9-0802-4065-9630-2a2200060db4","Type":"ContainerStarted","Data":"23cfb2637f3a883c7374c352adf87946c523400b25fa81403565d45084fa761d"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.493300 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-14bb-account-create-update-s96dt" event={"ID":"428ee377-ea6d-4897-890a-6676a318c95a","Type":"ContainerDied","Data":"63b00bee2eaf0f6558d377b5f684870d48e1d780311b068d69c9f62f2403da66"} Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.493375 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63b00bee2eaf0f6558d377b5f684870d48e1d780311b068d69c9f62f2403da66" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.493451 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-14bb-account-create-update-s96dt" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.588113 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-stqc7" podStartSLOduration=10.095467911 podStartE2EDuration="21.588088739s" podCreationTimestamp="2025-12-03 17:17:45 +0000 UTC" firstStartedPulling="2025-12-03 17:17:54.059694395 +0000 UTC m=+1329.261071266" lastFinishedPulling="2025-12-03 17:18:05.552315223 +0000 UTC m=+1340.753692094" observedRunningTime="2025-12-03 17:18:06.544756557 +0000 UTC m=+1341.746133428" watchObservedRunningTime="2025-12-03 17:18:06.588088739 +0000 UTC m=+1341.789465600" Dec 03 17:18:06 crc kubenswrapper[4758]: I1203 17:18:06.595038 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5czb6-config-2hn4j" podStartSLOduration=8.595014545 podStartE2EDuration="8.595014545s" podCreationTimestamp="2025-12-03 17:17:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:06.585503929 +0000 UTC m=+1341.786880800" watchObservedRunningTime="2025-12-03 17:18:06.595014545 +0000 UTC m=+1341.796391406" Dec 03 17:18:07 crc kubenswrapper[4758]: I1203 17:18:07.508624 4758 generic.go:334] "Generic (PLEG): container finished" podID="9ed875bd-af54-48cb-91ca-b6f92fda869f" containerID="6d70a91ca70352bcd228584cfc0a35eeeed5d96455696994c096ca68be4c414f" exitCode=0 Dec 03 17:18:07 crc kubenswrapper[4758]: I1203 17:18:07.508722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-2hn4j" event={"ID":"9ed875bd-af54-48cb-91ca-b6f92fda869f","Type":"ContainerDied","Data":"6d70a91ca70352bcd228584cfc0a35eeeed5d96455696994c096ca68be4c414f"} Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.024335 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037331 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-scripts\") pod \"9ed875bd-af54-48cb-91ca-b6f92fda869f\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037403 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run-ovn\") pod \"9ed875bd-af54-48cb-91ca-b6f92fda869f\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037471 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-log-ovn\") pod \"9ed875bd-af54-48cb-91ca-b6f92fda869f\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037532 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-additional-scripts\") pod \"9ed875bd-af54-48cb-91ca-b6f92fda869f\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037573 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9ed875bd-af54-48cb-91ca-b6f92fda869f" (UID: "9ed875bd-af54-48cb-91ca-b6f92fda869f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037598 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9ed875bd-af54-48cb-91ca-b6f92fda869f" (UID: "9ed875bd-af54-48cb-91ca-b6f92fda869f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037746 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frl42\" (UniqueName: \"kubernetes.io/projected/9ed875bd-af54-48cb-91ca-b6f92fda869f-kube-api-access-frl42\") pod \"9ed875bd-af54-48cb-91ca-b6f92fda869f\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037777 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run\") pod \"9ed875bd-af54-48cb-91ca-b6f92fda869f\" (UID: \"9ed875bd-af54-48cb-91ca-b6f92fda869f\") " Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.037848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run" (OuterVolumeSpecName: "var-run") pod "9ed875bd-af54-48cb-91ca-b6f92fda869f" (UID: "9ed875bd-af54-48cb-91ca-b6f92fda869f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.038476 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9ed875bd-af54-48cb-91ca-b6f92fda869f" (UID: "9ed875bd-af54-48cb-91ca-b6f92fda869f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.038843 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-scripts" (OuterVolumeSpecName: "scripts") pod "9ed875bd-af54-48cb-91ca-b6f92fda869f" (UID: "9ed875bd-af54-48cb-91ca-b6f92fda869f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.039185 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.039216 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.039228 4758 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.039241 4758 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed875bd-af54-48cb-91ca-b6f92fda869f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.039255 4758 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ed875bd-af54-48cb-91ca-b6f92fda869f-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.046261 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed875bd-af54-48cb-91ca-b6f92fda869f-kube-api-access-frl42" (OuterVolumeSpecName: "kube-api-access-frl42") pod "9ed875bd-af54-48cb-91ca-b6f92fda869f" (UID: "9ed875bd-af54-48cb-91ca-b6f92fda869f"). InnerVolumeSpecName "kube-api-access-frl42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.140827 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frl42\" (UniqueName: \"kubernetes.io/projected/9ed875bd-af54-48cb-91ca-b6f92fda869f-kube-api-access-frl42\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.546066 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-2hn4j" event={"ID":"9ed875bd-af54-48cb-91ca-b6f92fda869f","Type":"ContainerDied","Data":"8800a232afab58d607dcc7069076495a9e2397accbf7ddeb59dd8175f8f710d5"} Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.546124 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8800a232afab58d607dcc7069076495a9e2397accbf7ddeb59dd8175f8f710d5" Dec 03 17:18:10 crc kubenswrapper[4758]: I1203 17:18:10.546168 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-2hn4j" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.125861 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5czb6-config-2hn4j"] Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.125904 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5czb6-config-2hn4j"] Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.277187 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5czb6-config-6phz7"] Dec 03 17:18:11 crc kubenswrapper[4758]: E1203 17:18:11.277850 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed875bd-af54-48cb-91ca-b6f92fda869f" containerName="ovn-config" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.277881 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed875bd-af54-48cb-91ca-b6f92fda869f" containerName="ovn-config" Dec 03 17:18:11 crc kubenswrapper[4758]: E1203 17:18:11.277907 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20469a6a-4abc-4e0b-886f-ba199ee5caa3" containerName="mariadb-database-create" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.277916 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="20469a6a-4abc-4e0b-886f-ba199ee5caa3" containerName="mariadb-database-create" Dec 03 17:18:11 crc kubenswrapper[4758]: E1203 17:18:11.277959 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428ee377-ea6d-4897-890a-6676a318c95a" containerName="mariadb-account-create-update" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.277968 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="428ee377-ea6d-4897-890a-6676a318c95a" containerName="mariadb-account-create-update" Dec 03 17:18:11 crc kubenswrapper[4758]: E1203 17:18:11.277980 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde0a172-de16-4feb-a2e0-c51ccb8f41e2" containerName="mariadb-account-create-update" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.277988 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde0a172-de16-4feb-a2e0-c51ccb8f41e2" containerName="mariadb-account-create-update" Dec 03 17:18:11 crc kubenswrapper[4758]: E1203 17:18:11.278000 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a0e71a2-24ae-4842-935e-5ed7a9bc7145" containerName="mariadb-database-create" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.278007 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a0e71a2-24ae-4842-935e-5ed7a9bc7145" containerName="mariadb-database-create" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.278207 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="20469a6a-4abc-4e0b-886f-ba199ee5caa3" containerName="mariadb-database-create" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.278253 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde0a172-de16-4feb-a2e0-c51ccb8f41e2" containerName="mariadb-account-create-update" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.278270 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a0e71a2-24ae-4842-935e-5ed7a9bc7145" containerName="mariadb-database-create" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.278286 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="428ee377-ea6d-4897-890a-6676a318c95a" containerName="mariadb-account-create-update" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.278303 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed875bd-af54-48cb-91ca-b6f92fda869f" containerName="ovn-config" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.279092 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.282226 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.291421 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6-config-6phz7"] Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.394566 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.394661 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.464615 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-additional-scripts\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.465870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8ttv\" (UniqueName: \"kubernetes.io/projected/8a280581-fa8d-4c86-9df8-71c995158185-kube-api-access-d8ttv\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.465944 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-log-ovn\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.466022 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-scripts\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.466059 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run-ovn\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.466309 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568421 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-additional-scripts\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568482 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8ttv\" (UniqueName: \"kubernetes.io/projected/8a280581-fa8d-4c86-9df8-71c995158185-kube-api-access-d8ttv\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568520 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-log-ovn\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568561 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-scripts\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568591 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run-ovn\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568810 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run-ovn\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568823 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.568934 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-log-ovn\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.570024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-additional-scripts\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.574154 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-scripts\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.592801 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8ttv\" (UniqueName: \"kubernetes.io/projected/8a280581-fa8d-4c86-9df8-71c995158185-kube-api-access-d8ttv\") pod \"ovn-controller-5czb6-config-6phz7\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:11 crc kubenswrapper[4758]: I1203 17:18:11.598554 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:12 crc kubenswrapper[4758]: I1203 17:18:12.241415 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5czb6-config-6phz7"] Dec 03 17:18:12 crc kubenswrapper[4758]: W1203 17:18:12.251375 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a280581_fa8d_4c86_9df8_71c995158185.slice/crio-881731ebe95a114a06886c253bfbf743e4c3c96ec51682ab72a8f4a10cea80be WatchSource:0}: Error finding container 881731ebe95a114a06886c253bfbf743e4c3c96ec51682ab72a8f4a10cea80be: Status 404 returned error can't find the container with id 881731ebe95a114a06886c253bfbf743e4c3c96ec51682ab72a8f4a10cea80be Dec 03 17:18:12 crc kubenswrapper[4758]: I1203 17:18:12.579885 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06"} Dec 03 17:18:12 crc kubenswrapper[4758]: I1203 17:18:12.580339 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81"} Dec 03 17:18:12 crc kubenswrapper[4758]: I1203 17:18:12.582253 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qr4sd" event={"ID":"eb23ba13-deef-46a6-a81d-a223ad56024b","Type":"ContainerStarted","Data":"288ec7994c38909faa37c6fb97e7c615461c154f58fffb1db5fac5164c538a3c"} Dec 03 17:18:12 crc kubenswrapper[4758]: I1203 17:18:12.588962 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-6phz7" event={"ID":"8a280581-fa8d-4c86-9df8-71c995158185","Type":"ContainerStarted","Data":"881731ebe95a114a06886c253bfbf743e4c3c96ec51682ab72a8f4a10cea80be"} Dec 03 17:18:12 crc kubenswrapper[4758]: I1203 17:18:12.609267 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-qr4sd" podStartSLOduration=2.551738512 podStartE2EDuration="45.609242077s" podCreationTimestamp="2025-12-03 17:17:27 +0000 UTC" firstStartedPulling="2025-12-03 17:17:28.677752323 +0000 UTC m=+1303.879129184" lastFinishedPulling="2025-12-03 17:18:11.735255888 +0000 UTC m=+1346.936632749" observedRunningTime="2025-12-03 17:18:12.607019916 +0000 UTC m=+1347.808396777" watchObservedRunningTime="2025-12-03 17:18:12.609242077 +0000 UTC m=+1347.810618938" Dec 03 17:18:13 crc kubenswrapper[4758]: I1203 17:18:13.127294 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed875bd-af54-48cb-91ca-b6f92fda869f" path="/var/lib/kubelet/pods/9ed875bd-af54-48cb-91ca-b6f92fda869f/volumes" Dec 03 17:18:13 crc kubenswrapper[4758]: I1203 17:18:13.600210 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-6phz7" event={"ID":"8a280581-fa8d-4c86-9df8-71c995158185","Type":"ContainerStarted","Data":"6c1f749418d08e981d41d7fa033a9510b208034ee61c5bce425ac2db39489187"} Dec 03 17:18:15 crc kubenswrapper[4758]: I1203 17:18:15.630144 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f"} Dec 03 17:18:15 crc kubenswrapper[4758]: I1203 17:18:15.630904 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637"} Dec 03 17:18:15 crc kubenswrapper[4758]: I1203 17:18:15.635799 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a280581-fa8d-4c86-9df8-71c995158185" containerID="6c1f749418d08e981d41d7fa033a9510b208034ee61c5bce425ac2db39489187" exitCode=0 Dec 03 17:18:15 crc kubenswrapper[4758]: I1203 17:18:15.635845 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-6phz7" event={"ID":"8a280581-fa8d-4c86-9df8-71c995158185","Type":"ContainerDied","Data":"6c1f749418d08e981d41d7fa033a9510b208034ee61c5bce425ac2db39489187"} Dec 03 17:18:16 crc kubenswrapper[4758]: I1203 17:18:16.661436 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27"} Dec 03 17:18:16 crc kubenswrapper[4758]: I1203 17:18:16.661581 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422"} Dec 03 17:18:16 crc kubenswrapper[4758]: I1203 17:18:16.661594 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerStarted","Data":"1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e"} Dec 03 17:18:16 crc kubenswrapper[4758]: I1203 17:18:16.712506 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.870292647 podStartE2EDuration="1m17.712475479s" podCreationTimestamp="2025-12-03 17:16:59 +0000 UTC" firstStartedPulling="2025-12-03 17:17:33.673055675 +0000 UTC m=+1308.874432536" lastFinishedPulling="2025-12-03 17:18:11.515238507 +0000 UTC m=+1346.716615368" observedRunningTime="2025-12-03 17:18:16.696093837 +0000 UTC m=+1351.897470718" watchObservedRunningTime="2025-12-03 17:18:16.712475479 +0000 UTC m=+1351.913852340" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.053279 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-cq2zt"] Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.055647 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.057210 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.060449 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.070857 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-cq2zt"] Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198313 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-log-ovn\") pod \"8a280581-fa8d-4c86-9df8-71c995158185\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198455 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run\") pod \"8a280581-fa8d-4c86-9df8-71c995158185\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198479 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8a280581-fa8d-4c86-9df8-71c995158185" (UID: "8a280581-fa8d-4c86-9df8-71c995158185"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198514 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-scripts\") pod \"8a280581-fa8d-4c86-9df8-71c995158185\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198561 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run" (OuterVolumeSpecName: "var-run") pod "8a280581-fa8d-4c86-9df8-71c995158185" (UID: "8a280581-fa8d-4c86-9df8-71c995158185"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198554 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run-ovn\") pod \"8a280581-fa8d-4c86-9df8-71c995158185\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198658 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-additional-scripts\") pod \"8a280581-fa8d-4c86-9df8-71c995158185\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198728 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8ttv\" (UniqueName: \"kubernetes.io/projected/8a280581-fa8d-4c86-9df8-71c995158185-kube-api-access-d8ttv\") pod \"8a280581-fa8d-4c86-9df8-71c995158185\" (UID: \"8a280581-fa8d-4c86-9df8-71c995158185\") " Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.198801 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8a280581-fa8d-4c86-9df8-71c995158185" (UID: "8a280581-fa8d-4c86-9df8-71c995158185"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199057 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-config\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199105 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mflxd\" (UniqueName: \"kubernetes.io/projected/5a116d59-0671-4da9-8462-414abf4d5cba-kube-api-access-mflxd\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199364 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199741 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8a280581-fa8d-4c86-9df8-71c995158185" (UID: "8a280581-fa8d-4c86-9df8-71c995158185"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199797 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199885 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.199970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-scripts" (OuterVolumeSpecName: "scripts") pod "8a280581-fa8d-4c86-9df8-71c995158185" (UID: "8a280581-fa8d-4c86-9df8-71c995158185"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.200051 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.200073 4758 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.200085 4758 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.200101 4758 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a280581-fa8d-4c86-9df8-71c995158185-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.205949 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a280581-fa8d-4c86-9df8-71c995158185-kube-api-access-d8ttv" (OuterVolumeSpecName: "kube-api-access-d8ttv") pod "8a280581-fa8d-4c86-9df8-71c995158185" (UID: "8a280581-fa8d-4c86-9df8-71c995158185"). InnerVolumeSpecName "kube-api-access-d8ttv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.301719 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.301833 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-config\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.301887 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mflxd\" (UniqueName: \"kubernetes.io/projected/5a116d59-0671-4da9-8462-414abf4d5cba-kube-api-access-mflxd\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.301962 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.302049 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.302094 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.302176 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8ttv\" (UniqueName: \"kubernetes.io/projected/8a280581-fa8d-4c86-9df8-71c995158185-kube-api-access-d8ttv\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.302194 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a280581-fa8d-4c86-9df8-71c995158185-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.303501 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.303545 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.303563 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-config\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.303531 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.303805 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.325831 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mflxd\" (UniqueName: \"kubernetes.io/projected/5a116d59-0671-4da9-8462-414abf4d5cba-kube-api-access-mflxd\") pod \"dnsmasq-dns-5c79d794d7-cq2zt\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.390521 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.673709 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6-config-6phz7" event={"ID":"8a280581-fa8d-4c86-9df8-71c995158185","Type":"ContainerDied","Data":"881731ebe95a114a06886c253bfbf743e4c3c96ec51682ab72a8f4a10cea80be"} Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.674189 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="881731ebe95a114a06886c253bfbf743e4c3c96ec51682ab72a8f4a10cea80be" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.673808 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6-config-6phz7" Dec 03 17:18:17 crc kubenswrapper[4758]: I1203 17:18:17.896228 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-cq2zt"] Dec 03 17:18:17 crc kubenswrapper[4758]: W1203 17:18:17.901915 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a116d59_0671_4da9_8462_414abf4d5cba.slice/crio-2b515d087e7119667d97a948c38038928bd02cdfa0a7861451972508bbdb0c28 WatchSource:0}: Error finding container 2b515d087e7119667d97a948c38038928bd02cdfa0a7861451972508bbdb0c28: Status 404 returned error can't find the container with id 2b515d087e7119667d97a948c38038928bd02cdfa0a7861451972508bbdb0c28 Dec 03 17:18:18 crc kubenswrapper[4758]: I1203 17:18:18.207369 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5czb6-config-6phz7"] Dec 03 17:18:18 crc kubenswrapper[4758]: I1203 17:18:18.224067 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5czb6-config-6phz7"] Dec 03 17:18:18 crc kubenswrapper[4758]: I1203 17:18:18.685944 4758 generic.go:334] "Generic (PLEG): container finished" podID="5a116d59-0671-4da9-8462-414abf4d5cba" containerID="9d7046ea95e640d0086ac42f04e209cffa03ce4246b26726886996b2150de591" exitCode=0 Dec 03 17:18:18 crc kubenswrapper[4758]: I1203 17:18:18.686011 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" event={"ID":"5a116d59-0671-4da9-8462-414abf4d5cba","Type":"ContainerDied","Data":"9d7046ea95e640d0086ac42f04e209cffa03ce4246b26726886996b2150de591"} Dec 03 17:18:18 crc kubenswrapper[4758]: I1203 17:18:18.686051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" event={"ID":"5a116d59-0671-4da9-8462-414abf4d5cba","Type":"ContainerStarted","Data":"2b515d087e7119667d97a948c38038928bd02cdfa0a7861451972508bbdb0c28"} Dec 03 17:18:19 crc kubenswrapper[4758]: I1203 17:18:19.126171 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a280581-fa8d-4c86-9df8-71c995158185" path="/var/lib/kubelet/pods/8a280581-fa8d-4c86-9df8-71c995158185/volumes" Dec 03 17:18:19 crc kubenswrapper[4758]: I1203 17:18:19.702059 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" event={"ID":"5a116d59-0671-4da9-8462-414abf4d5cba","Type":"ContainerStarted","Data":"488f1d0a36296d211053283862402de8f3c6cf3850cd27b34d62ad8956fc53d4"} Dec 03 17:18:19 crc kubenswrapper[4758]: I1203 17:18:19.702257 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:19 crc kubenswrapper[4758]: I1203 17:18:19.723459 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" podStartSLOduration=2.723437199 podStartE2EDuration="2.723437199s" podCreationTimestamp="2025-12-03 17:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:19.722866294 +0000 UTC m=+1354.924243165" watchObservedRunningTime="2025-12-03 17:18:19.723437199 +0000 UTC m=+1354.924814060" Dec 03 17:18:23 crc kubenswrapper[4758]: I1203 17:18:23.749714 4758 generic.go:334] "Generic (PLEG): container finished" podID="8e2e52f9-0802-4065-9630-2a2200060db4" containerID="23cfb2637f3a883c7374c352adf87946c523400b25fa81403565d45084fa761d" exitCode=0 Dec 03 17:18:23 crc kubenswrapper[4758]: I1203 17:18:23.749800 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stqc7" event={"ID":"8e2e52f9-0802-4065-9630-2a2200060db4","Type":"ContainerDied","Data":"23cfb2637f3a883c7374c352adf87946c523400b25fa81403565d45084fa761d"} Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.127041 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stqc7" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.281973 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j946t\" (UniqueName: \"kubernetes.io/projected/8e2e52f9-0802-4065-9630-2a2200060db4-kube-api-access-j946t\") pod \"8e2e52f9-0802-4065-9630-2a2200060db4\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.282094 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-combined-ca-bundle\") pod \"8e2e52f9-0802-4065-9630-2a2200060db4\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.282210 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-config-data\") pod \"8e2e52f9-0802-4065-9630-2a2200060db4\" (UID: \"8e2e52f9-0802-4065-9630-2a2200060db4\") " Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.292194 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2e52f9-0802-4065-9630-2a2200060db4-kube-api-access-j946t" (OuterVolumeSpecName: "kube-api-access-j946t") pod "8e2e52f9-0802-4065-9630-2a2200060db4" (UID: "8e2e52f9-0802-4065-9630-2a2200060db4"). InnerVolumeSpecName "kube-api-access-j946t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.326383 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e2e52f9-0802-4065-9630-2a2200060db4" (UID: "8e2e52f9-0802-4065-9630-2a2200060db4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.339937 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-config-data" (OuterVolumeSpecName: "config-data") pod "8e2e52f9-0802-4065-9630-2a2200060db4" (UID: "8e2e52f9-0802-4065-9630-2a2200060db4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.385431 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.385475 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j946t\" (UniqueName: \"kubernetes.io/projected/8e2e52f9-0802-4065-9630-2a2200060db4-kube-api-access-j946t\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.385490 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e2e52f9-0802-4065-9630-2a2200060db4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.771592 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stqc7" event={"ID":"8e2e52f9-0802-4065-9630-2a2200060db4","Type":"ContainerDied","Data":"7914c3f813ccfdc70f5f97d8ec44ee6e028ad454a9df801183820f2772960f30"} Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.772110 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7914c3f813ccfdc70f5f97d8ec44ee6e028ad454a9df801183820f2772960f30" Dec 03 17:18:25 crc kubenswrapper[4758]: I1203 17:18:25.771744 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stqc7" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.123816 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mhwrm"] Dec 03 17:18:26 crc kubenswrapper[4758]: E1203 17:18:26.124552 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2e52f9-0802-4065-9630-2a2200060db4" containerName="keystone-db-sync" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.124577 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2e52f9-0802-4065-9630-2a2200060db4" containerName="keystone-db-sync" Dec 03 17:18:26 crc kubenswrapper[4758]: E1203 17:18:26.124621 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a280581-fa8d-4c86-9df8-71c995158185" containerName="ovn-config" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.124630 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a280581-fa8d-4c86-9df8-71c995158185" containerName="ovn-config" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.124933 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a280581-fa8d-4c86-9df8-71c995158185" containerName="ovn-config" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.124962 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2e52f9-0802-4065-9630-2a2200060db4" containerName="keystone-db-sync" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.125965 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.145523 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-cq2zt"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.146011 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" containerName="dnsmasq-dns" containerID="cri-o://488f1d0a36296d211053283862402de8f3c6cf3850cd27b34d62ad8956fc53d4" gracePeriod=10 Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.148371 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.148990 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.150092 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9h99t" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.150932 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.153339 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.174079 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.221702 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-fernet-keys\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.221773 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzgkj\" (UniqueName: \"kubernetes.io/projected/4718597a-e89e-4b71-9f09-a559c94cd69d-kube-api-access-fzgkj\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.221985 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-config-data\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.222039 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-combined-ca-bundle\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.222170 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-scripts\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.222293 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-credential-keys\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.328907 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mhwrm"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.337609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-config-data\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.337673 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-combined-ca-bundle\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.337751 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-scripts\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.337806 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-credential-keys\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.337836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-fernet-keys\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.337872 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzgkj\" (UniqueName: \"kubernetes.io/projected/4718597a-e89e-4b71-9f09-a559c94cd69d-kube-api-access-fzgkj\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.379395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-combined-ca-bundle\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.393050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-scripts\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.399643 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-bzchc"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.404577 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzgkj\" (UniqueName: \"kubernetes.io/projected/4718597a-e89e-4b71-9f09-a559c94cd69d-kube-api-access-fzgkj\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.407613 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-config-data\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.411417 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.416550 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-fernet-keys\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.429285 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-credential-keys\") pod \"keystone-bootstrap-mhwrm\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.432875 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-bzchc"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.552302 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.553315 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq8c9\" (UniqueName: \"kubernetes.io/projected/2567e8e3-84ee-4477-94c8-c61b02086f57-kube-api-access-zq8c9\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.553588 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-config\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.553721 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-svc\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.554016 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.554140 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.554248 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.595848 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-j7q7w"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.597448 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.615729 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.618132 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.633106 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.634053 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.634387 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tt8c6" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.634835 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.645786 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.657498 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-svc\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.657608 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-config\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.657650 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.657707 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.657745 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.657920 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjcm\" (UniqueName: \"kubernetes.io/projected/e6bb6321-e853-40ce-a5f6-7b7a53e26595-kube-api-access-2tjcm\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.658011 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq8c9\" (UniqueName: \"kubernetes.io/projected/2567e8e3-84ee-4477-94c8-c61b02086f57-kube-api-access-zq8c9\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.658095 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-combined-ca-bundle\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.658169 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-config\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.658925 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.659376 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-config\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.660031 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-svc\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.670185 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-j7q7w"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.681832 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-6l8jp"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.687340 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.688587 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.693590 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.705603 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.705919 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.706015 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5fmvm" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.688738 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.765161 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-combined-ca-bundle\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.765320 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-scripts\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.765477 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbfaa3a1-7e61-4537-b953-07269f386adc-etc-machine-id\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.765585 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cl5m\" (UniqueName: \"kubernetes.io/projected/dbfaa3a1-7e61-4537-b953-07269f386adc-kube-api-access-2cl5m\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766126 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-config\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766212 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-config-data\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766245 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-combined-ca-bundle\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-run-httpd\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766427 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj57t\" (UniqueName: \"kubernetes.io/projected/994006af-c89c-42b1-b0b7-f6a247e83cab-kube-api-access-sj57t\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766460 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-scripts\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766493 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766607 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-db-sync-config-data\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766669 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjcm\" (UniqueName: \"kubernetes.io/projected/e6bb6321-e853-40ce-a5f6-7b7a53e26595-kube-api-access-2tjcm\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766771 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-log-httpd\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.766849 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-config-data\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.795631 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-combined-ca-bundle\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.797647 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-config\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.810565 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6l8jp"] Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.830715 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq8c9\" (UniqueName: \"kubernetes.io/projected/2567e8e3-84ee-4477-94c8-c61b02086f57-kube-api-access-zq8c9\") pod \"dnsmasq-dns-5b868669f-bzchc\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.842579 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjcm\" (UniqueName: \"kubernetes.io/projected/e6bb6321-e853-40ce-a5f6-7b7a53e26595-kube-api-access-2tjcm\") pod \"neutron-db-sync-j7q7w\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.881380 4758 generic.go:334] "Generic (PLEG): container finished" podID="5a116d59-0671-4da9-8462-414abf4d5cba" containerID="488f1d0a36296d211053283862402de8f3c6cf3850cd27b34d62ad8956fc53d4" exitCode=0 Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.881902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" event={"ID":"5a116d59-0671-4da9-8462-414abf4d5cba","Type":"ContainerDied","Data":"488f1d0a36296d211053283862402de8f3c6cf3850cd27b34d62ad8956fc53d4"} Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-config-data\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887515 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-combined-ca-bundle\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887559 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-run-httpd\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887603 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887648 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj57t\" (UniqueName: \"kubernetes.io/projected/994006af-c89c-42b1-b0b7-f6a247e83cab-kube-api-access-sj57t\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887674 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-scripts\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887723 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887787 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-db-sync-config-data\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887809 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-log-httpd\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887843 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-config-data\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887902 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-scripts\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.887951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbfaa3a1-7e61-4537-b953-07269f386adc-etc-machine-id\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.888000 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cl5m\" (UniqueName: \"kubernetes.io/projected/dbfaa3a1-7e61-4537-b953-07269f386adc-kube-api-access-2cl5m\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.888220 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-run-httpd\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.906905 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbfaa3a1-7e61-4537-b953-07269f386adc-etc-machine-id\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.907575 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.908348 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-log-httpd\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.915372 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.916243 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.921924 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-scripts\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.923346 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-config-data\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.932896 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-db-sync-config-data\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.941413 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-scripts\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.943278 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-combined-ca-bundle\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.951835 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-config-data\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.952446 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cl5m\" (UniqueName: \"kubernetes.io/projected/dbfaa3a1-7e61-4537-b953-07269f386adc-kube-api-access-2cl5m\") pod \"cinder-db-sync-6l8jp\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:26 crc kubenswrapper[4758]: I1203 17:18:26.984172 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj57t\" (UniqueName: \"kubernetes.io/projected/994006af-c89c-42b1-b0b7-f6a247e83cab-kube-api-access-sj57t\") pod \"ceilometer-0\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " pod="openstack/ceilometer-0" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.060107 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-q2vlf"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.061665 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.068195 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-585nf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.068757 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.105918 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-q2vlf"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.117167 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.144408 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-bzchc"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.144456 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4jmgb"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.145711 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.147428 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.156425 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.156812 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.157005 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-87xxm" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.157135 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4jmgb"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.165924 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4ggvx"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.168160 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.178759 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4ggvx"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.198117 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.215855 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-db-sync-config-data\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.215904 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-combined-ca-bundle\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.216016 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/502df21c-0370-40b8-b3ef-0e55992b3008-logs\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.216039 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wzts\" (UniqueName: \"kubernetes.io/projected/6b98f774-a8d0-460e-a718-c119dbd1f203-kube-api-access-7wzts\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.216071 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-config-data\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.216086 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc7lk\" (UniqueName: \"kubernetes.io/projected/502df21c-0370-40b8-b3ef-0e55992b3008-kube-api-access-gc7lk\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.216108 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-combined-ca-bundle\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.216173 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-scripts\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.297489 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318053 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4czhn\" (UniqueName: \"kubernetes.io/projected/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-kube-api-access-4czhn\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318160 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-config\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318198 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/502df21c-0370-40b8-b3ef-0e55992b3008-logs\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318236 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318264 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wzts\" (UniqueName: \"kubernetes.io/projected/6b98f774-a8d0-460e-a718-c119dbd1f203-kube-api-access-7wzts\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318307 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-config-data\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318326 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc7lk\" (UniqueName: \"kubernetes.io/projected/502df21c-0370-40b8-b3ef-0e55992b3008-kube-api-access-gc7lk\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318356 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-combined-ca-bundle\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318407 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318480 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-scripts\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318507 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318550 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-db-sync-config-data\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.318584 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-combined-ca-bundle\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.320845 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/502df21c-0370-40b8-b3ef-0e55992b3008-logs\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.329704 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-db-sync-config-data\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.329753 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-combined-ca-bundle\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.331507 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-scripts\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.333056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-config-data\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.333541 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-combined-ca-bundle\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.365776 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wzts\" (UniqueName: \"kubernetes.io/projected/6b98f774-a8d0-460e-a718-c119dbd1f203-kube-api-access-7wzts\") pod \"barbican-db-sync-q2vlf\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.374329 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc7lk\" (UniqueName: \"kubernetes.io/projected/502df21c-0370-40b8-b3ef-0e55992b3008-kube-api-access-gc7lk\") pod \"placement-db-sync-4jmgb\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.423752 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.423909 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-nb\") pod \"5a116d59-0671-4da9-8462-414abf4d5cba\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.424085 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mflxd\" (UniqueName: \"kubernetes.io/projected/5a116d59-0671-4da9-8462-414abf4d5cba-kube-api-access-mflxd\") pod \"5a116d59-0671-4da9-8462-414abf4d5cba\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.424137 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-svc\") pod \"5a116d59-0671-4da9-8462-414abf4d5cba\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.424181 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-sb\") pod \"5a116d59-0671-4da9-8462-414abf4d5cba\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.424310 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-swift-storage-0\") pod \"5a116d59-0671-4da9-8462-414abf4d5cba\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.424477 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-config\") pod \"5a116d59-0671-4da9-8462-414abf4d5cba\" (UID: \"5a116d59-0671-4da9-8462-414abf4d5cba\") " Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.425207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.425358 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.425413 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.425472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.425610 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4czhn\" (UniqueName: \"kubernetes.io/projected/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-kube-api-access-4czhn\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.429417 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-config\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.431143 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-config\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.433253 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.434357 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.435194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.454390 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.475359 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4czhn\" (UniqueName: \"kubernetes.io/projected/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-kube-api-access-4czhn\") pod \"dnsmasq-dns-cf78879c9-4ggvx\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.489829 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a116d59-0671-4da9-8462-414abf4d5cba-kube-api-access-mflxd" (OuterVolumeSpecName: "kube-api-access-mflxd") pod "5a116d59-0671-4da9-8462-414abf4d5cba" (UID: "5a116d59-0671-4da9-8462-414abf4d5cba"). InnerVolumeSpecName "kube-api-access-mflxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.509170 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4jmgb" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.575500 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5a116d59-0671-4da9-8462-414abf4d5cba" (UID: "5a116d59-0671-4da9-8462-414abf4d5cba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.577481 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mflxd\" (UniqueName: \"kubernetes.io/projected/5a116d59-0671-4da9-8462-414abf4d5cba-kube-api-access-mflxd\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.577580 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.581553 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.612225 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5a116d59-0671-4da9-8462-414abf4d5cba" (UID: "5a116d59-0671-4da9-8462-414abf4d5cba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.613443 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5a116d59-0671-4da9-8462-414abf4d5cba" (UID: "5a116d59-0671-4da9-8462-414abf4d5cba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.619991 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-config" (OuterVolumeSpecName: "config") pod "5a116d59-0671-4da9-8462-414abf4d5cba" (UID: "5a116d59-0671-4da9-8462-414abf4d5cba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.637983 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a116d59-0671-4da9-8462-414abf4d5cba" (UID: "5a116d59-0671-4da9-8462-414abf4d5cba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.643131 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mhwrm"] Dec 03 17:18:27 crc kubenswrapper[4758]: W1203 17:18:27.648210 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4718597a_e89e_4b71_9f09_a559c94cd69d.slice/crio-3a42ea0307f61df76188ddc87516ba4ed62d1bfc7ab47d162b8503e53767a469 WatchSource:0}: Error finding container 3a42ea0307f61df76188ddc87516ba4ed62d1bfc7ab47d162b8503e53767a469: Status 404 returned error can't find the container with id 3a42ea0307f61df76188ddc87516ba4ed62d1bfc7ab47d162b8503e53767a469 Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.680257 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.680328 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.680345 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.680389 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a116d59-0671-4da9-8462-414abf4d5cba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.809165 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-bzchc"] Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.905338 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-bzchc" event={"ID":"2567e8e3-84ee-4477-94c8-c61b02086f57","Type":"ContainerStarted","Data":"62610ba25a60f1462ad0a30a1aa2b204abb23c328067bb6d5671975367231c26"} Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.908605 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mhwrm" event={"ID":"4718597a-e89e-4b71-9f09-a559c94cd69d","Type":"ContainerStarted","Data":"3a42ea0307f61df76188ddc87516ba4ed62d1bfc7ab47d162b8503e53767a469"} Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.912279 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" event={"ID":"5a116d59-0671-4da9-8462-414abf4d5cba","Type":"ContainerDied","Data":"2b515d087e7119667d97a948c38038928bd02cdfa0a7861451972508bbdb0c28"} Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.912358 4758 scope.go:117] "RemoveContainer" containerID="488f1d0a36296d211053283862402de8f3c6cf3850cd27b34d62ad8956fc53d4" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.912402 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-cq2zt" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.945575 4758 scope.go:117] "RemoveContainer" containerID="9d7046ea95e640d0086ac42f04e209cffa03ce4246b26726886996b2150de591" Dec 03 17:18:27 crc kubenswrapper[4758]: I1203 17:18:27.982814 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-cq2zt"] Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.000038 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-cq2zt"] Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.069516 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6l8jp"] Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.077108 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-j7q7w"] Dec 03 17:18:28 crc kubenswrapper[4758]: W1203 17:18:28.083218 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbfaa3a1_7e61_4537_b953_07269f386adc.slice/crio-fce8c67cc946cbd7a2cbffe68752bb07c36640b8f446577a94c9f496ea5c28df WatchSource:0}: Error finding container fce8c67cc946cbd7a2cbffe68752bb07c36640b8f446577a94c9f496ea5c28df: Status 404 returned error can't find the container with id fce8c67cc946cbd7a2cbffe68752bb07c36640b8f446577a94c9f496ea5c28df Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.085360 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:18:28 crc kubenswrapper[4758]: W1203 17:18:28.086275 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6bb6321_e853_40ce_a5f6_7b7a53e26595.slice/crio-36cc319da74db7336be5004041fcbc892adc19c3b44896ceb1971c901ccccbcb WatchSource:0}: Error finding container 36cc319da74db7336be5004041fcbc892adc19c3b44896ceb1971c901ccccbcb: Status 404 returned error can't find the container with id 36cc319da74db7336be5004041fcbc892adc19c3b44896ceb1971c901ccccbcb Dec 03 17:18:28 crc kubenswrapper[4758]: W1203 17:18:28.088073 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod994006af_c89c_42b1_b0b7_f6a247e83cab.slice/crio-5933d3899283877952151052a4c96938c455a4b0a66e63b9e69fea0440546262 WatchSource:0}: Error finding container 5933d3899283877952151052a4c96938c455a4b0a66e63b9e69fea0440546262: Status 404 returned error can't find the container with id 5933d3899283877952151052a4c96938c455a4b0a66e63b9e69fea0440546262 Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.272580 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-q2vlf"] Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.341928 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4jmgb"] Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.378641 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4ggvx"] Dec 03 17:18:28 crc kubenswrapper[4758]: W1203 17:18:28.389482 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f4ec3d3_b3e2_426d_84bb_a4453d8d58a9.slice/crio-ffec55692f70664f94aed85a16b45310194554b9829fa8e481a713dfdfa9d505 WatchSource:0}: Error finding container ffec55692f70664f94aed85a16b45310194554b9829fa8e481a713dfdfa9d505: Status 404 returned error can't find the container with id ffec55692f70664f94aed85a16b45310194554b9829fa8e481a713dfdfa9d505 Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.925870 4758 generic.go:334] "Generic (PLEG): container finished" podID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerID="7b60c04c03cbdf8b0711c20b25062dc690eb658b38baf1bf3b829e0decd6d56d" exitCode=0 Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.926055 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" event={"ID":"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9","Type":"ContainerDied","Data":"7b60c04c03cbdf8b0711c20b25062dc690eb658b38baf1bf3b829e0decd6d56d"} Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.926413 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" event={"ID":"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9","Type":"ContainerStarted","Data":"ffec55692f70664f94aed85a16b45310194554b9829fa8e481a713dfdfa9d505"} Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.932332 4758 generic.go:334] "Generic (PLEG): container finished" podID="2567e8e3-84ee-4477-94c8-c61b02086f57" containerID="e784c9ddbfb6a964347f50c113014a8ccbd0eb0ce4d5853f73279cbbfd3db3c2" exitCode=0 Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.932407 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-bzchc" event={"ID":"2567e8e3-84ee-4477-94c8-c61b02086f57","Type":"ContainerDied","Data":"e784c9ddbfb6a964347f50c113014a8ccbd0eb0ce4d5853f73279cbbfd3db3c2"} Dec 03 17:18:28 crc kubenswrapper[4758]: I1203 17:18:28.998253 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-q2vlf" event={"ID":"6b98f774-a8d0-460e-a718-c119dbd1f203","Type":"ContainerStarted","Data":"6c16a1550f4aaad99325521243704066ddeea7e997248714a31a521d276af136"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.018437 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mhwrm" event={"ID":"4718597a-e89e-4b71-9f09-a559c94cd69d","Type":"ContainerStarted","Data":"7d1a2b7de5277b2da6cd394b4671007ed245d4da0dcaddd9b381996e33f4f2e5"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.071307 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerStarted","Data":"5933d3899283877952151052a4c96938c455a4b0a66e63b9e69fea0440546262"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.074410 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4jmgb" event={"ID":"502df21c-0370-40b8-b3ef-0e55992b3008","Type":"ContainerStarted","Data":"550829c342617ab5ac095f4de671c710c04dbc81e2d2657d218df85ed1fa8f5f"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.077244 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6l8jp" event={"ID":"dbfaa3a1-7e61-4537-b953-07269f386adc","Type":"ContainerStarted","Data":"fce8c67cc946cbd7a2cbffe68752bb07c36640b8f446577a94c9f496ea5c28df"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.082380 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j7q7w" event={"ID":"e6bb6321-e853-40ce-a5f6-7b7a53e26595","Type":"ContainerStarted","Data":"7247d58e71f563549d53c42917cec0911f2f98ae88698a1caa308986151490b6"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.082442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j7q7w" event={"ID":"e6bb6321-e853-40ce-a5f6-7b7a53e26595","Type":"ContainerStarted","Data":"36cc319da74db7336be5004041fcbc892adc19c3b44896ceb1971c901ccccbcb"} Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.098877 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mhwrm" podStartSLOduration=3.098838635 podStartE2EDuration="3.098838635s" podCreationTimestamp="2025-12-03 17:18:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:29.083164511 +0000 UTC m=+1364.284541382" watchObservedRunningTime="2025-12-03 17:18:29.098838635 +0000 UTC m=+1364.300215506" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.118550 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-j7q7w" podStartSLOduration=3.118527907 podStartE2EDuration="3.118527907s" podCreationTimestamp="2025-12-03 17:18:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:29.115104025 +0000 UTC m=+1364.316480886" watchObservedRunningTime="2025-12-03 17:18:29.118527907 +0000 UTC m=+1364.319904768" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.158654 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" path="/var/lib/kubelet/pods/5a116d59-0671-4da9-8462-414abf4d5cba/volumes" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.543748 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.589468 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.658754 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-config\") pod \"2567e8e3-84ee-4477-94c8-c61b02086f57\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.659237 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq8c9\" (UniqueName: \"kubernetes.io/projected/2567e8e3-84ee-4477-94c8-c61b02086f57-kube-api-access-zq8c9\") pod \"2567e8e3-84ee-4477-94c8-c61b02086f57\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.660196 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-sb\") pod \"2567e8e3-84ee-4477-94c8-c61b02086f57\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.660280 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-svc\") pod \"2567e8e3-84ee-4477-94c8-c61b02086f57\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.660303 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-swift-storage-0\") pod \"2567e8e3-84ee-4477-94c8-c61b02086f57\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.660428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-nb\") pod \"2567e8e3-84ee-4477-94c8-c61b02086f57\" (UID: \"2567e8e3-84ee-4477-94c8-c61b02086f57\") " Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.679500 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2567e8e3-84ee-4477-94c8-c61b02086f57-kube-api-access-zq8c9" (OuterVolumeSpecName: "kube-api-access-zq8c9") pod "2567e8e3-84ee-4477-94c8-c61b02086f57" (UID: "2567e8e3-84ee-4477-94c8-c61b02086f57"). InnerVolumeSpecName "kube-api-access-zq8c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.693126 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-config" (OuterVolumeSpecName: "config") pod "2567e8e3-84ee-4477-94c8-c61b02086f57" (UID: "2567e8e3-84ee-4477-94c8-c61b02086f57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.698152 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2567e8e3-84ee-4477-94c8-c61b02086f57" (UID: "2567e8e3-84ee-4477-94c8-c61b02086f57"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.706271 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2567e8e3-84ee-4477-94c8-c61b02086f57" (UID: "2567e8e3-84ee-4477-94c8-c61b02086f57"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.706393 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2567e8e3-84ee-4477-94c8-c61b02086f57" (UID: "2567e8e3-84ee-4477-94c8-c61b02086f57"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.714093 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2567e8e3-84ee-4477-94c8-c61b02086f57" (UID: "2567e8e3-84ee-4477-94c8-c61b02086f57"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.764470 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.764540 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq8c9\" (UniqueName: \"kubernetes.io/projected/2567e8e3-84ee-4477-94c8-c61b02086f57-kube-api-access-zq8c9\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.764561 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.764573 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.764584 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:29 crc kubenswrapper[4758]: I1203 17:18:29.764594 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2567e8e3-84ee-4477-94c8-c61b02086f57-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.099312 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-bzchc" event={"ID":"2567e8e3-84ee-4477-94c8-c61b02086f57","Type":"ContainerDied","Data":"62610ba25a60f1462ad0a30a1aa2b204abb23c328067bb6d5671975367231c26"} Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.099372 4758 scope.go:117] "RemoveContainer" containerID="e784c9ddbfb6a964347f50c113014a8ccbd0eb0ce4d5853f73279cbbfd3db3c2" Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.099498 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-bzchc" Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.127010 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" event={"ID":"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9","Type":"ContainerStarted","Data":"753c1f65ee759c0fdb55ace0c17c7d87b84db76d9c268e39c820bbac1853d21b"} Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.127386 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.209792 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-bzchc"] Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.259128 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-bzchc"] Dec 03 17:18:30 crc kubenswrapper[4758]: I1203 17:18:30.273951 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" podStartSLOduration=3.273911876 podStartE2EDuration="3.273911876s" podCreationTimestamp="2025-12-03 17:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:30.215719974 +0000 UTC m=+1365.417096855" watchObservedRunningTime="2025-12-03 17:18:30.273911876 +0000 UTC m=+1365.475288737" Dec 03 17:18:31 crc kubenswrapper[4758]: I1203 17:18:31.129552 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2567e8e3-84ee-4477-94c8-c61b02086f57" path="/var/lib/kubelet/pods/2567e8e3-84ee-4477-94c8-c61b02086f57/volumes" Dec 03 17:18:33 crc kubenswrapper[4758]: I1203 17:18:33.181803 4758 generic.go:334] "Generic (PLEG): container finished" podID="eb23ba13-deef-46a6-a81d-a223ad56024b" containerID="288ec7994c38909faa37c6fb97e7c615461c154f58fffb1db5fac5164c538a3c" exitCode=0 Dec 03 17:18:33 crc kubenswrapper[4758]: I1203 17:18:33.182235 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qr4sd" event={"ID":"eb23ba13-deef-46a6-a81d-a223ad56024b","Type":"ContainerDied","Data":"288ec7994c38909faa37c6fb97e7c615461c154f58fffb1db5fac5164c538a3c"} Dec 03 17:18:34 crc kubenswrapper[4758]: I1203 17:18:34.200072 4758 generic.go:334] "Generic (PLEG): container finished" podID="4718597a-e89e-4b71-9f09-a559c94cd69d" containerID="7d1a2b7de5277b2da6cd394b4671007ed245d4da0dcaddd9b381996e33f4f2e5" exitCode=0 Dec 03 17:18:34 crc kubenswrapper[4758]: I1203 17:18:34.200145 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mhwrm" event={"ID":"4718597a-e89e-4b71-9f09-a559c94cd69d","Type":"ContainerDied","Data":"7d1a2b7de5277b2da6cd394b4671007ed245d4da0dcaddd9b381996e33f4f2e5"} Dec 03 17:18:37 crc kubenswrapper[4758]: I1203 17:18:37.580035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:18:37 crc kubenswrapper[4758]: I1203 17:18:37.704171 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-zpczk"] Dec 03 17:18:37 crc kubenswrapper[4758]: I1203 17:18:37.704533 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" containerID="cri-o://e2b441905ca3a543bb1a2a712ed67ede3cc85561f57b550f7e41f589e4531e8a" gracePeriod=10 Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.273622 4758 generic.go:334] "Generic (PLEG): container finished" podID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerID="e2b441905ca3a543bb1a2a712ed67ede3cc85561f57b550f7e41f589e4531e8a" exitCode=0 Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.273700 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" event={"ID":"566c8d53-04a9-4937-bdab-4cac68a6ca49","Type":"ContainerDied","Data":"e2b441905ca3a543bb1a2a712ed67ede3cc85561f57b550f7e41f589e4531e8a"} Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.465577 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.581890 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-fernet-keys\") pod \"4718597a-e89e-4b71-9f09-a559c94cd69d\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.582064 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-config-data\") pod \"4718597a-e89e-4b71-9f09-a559c94cd69d\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.582100 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzgkj\" (UniqueName: \"kubernetes.io/projected/4718597a-e89e-4b71-9f09-a559c94cd69d-kube-api-access-fzgkj\") pod \"4718597a-e89e-4b71-9f09-a559c94cd69d\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.582317 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-scripts\") pod \"4718597a-e89e-4b71-9f09-a559c94cd69d\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.582383 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-combined-ca-bundle\") pod \"4718597a-e89e-4b71-9f09-a559c94cd69d\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.582469 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-credential-keys\") pod \"4718597a-e89e-4b71-9f09-a559c94cd69d\" (UID: \"4718597a-e89e-4b71-9f09-a559c94cd69d\") " Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.593481 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4718597a-e89e-4b71-9f09-a559c94cd69d" (UID: "4718597a-e89e-4b71-9f09-a559c94cd69d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.594052 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4718597a-e89e-4b71-9f09-a559c94cd69d-kube-api-access-fzgkj" (OuterVolumeSpecName: "kube-api-access-fzgkj") pod "4718597a-e89e-4b71-9f09-a559c94cd69d" (UID: "4718597a-e89e-4b71-9f09-a559c94cd69d"). InnerVolumeSpecName "kube-api-access-fzgkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.596622 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-scripts" (OuterVolumeSpecName: "scripts") pod "4718597a-e89e-4b71-9f09-a559c94cd69d" (UID: "4718597a-e89e-4b71-9f09-a559c94cd69d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.609055 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4718597a-e89e-4b71-9f09-a559c94cd69d" (UID: "4718597a-e89e-4b71-9f09-a559c94cd69d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.625924 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4718597a-e89e-4b71-9f09-a559c94cd69d" (UID: "4718597a-e89e-4b71-9f09-a559c94cd69d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.633789 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-config-data" (OuterVolumeSpecName: "config-data") pod "4718597a-e89e-4b71-9f09-a559c94cd69d" (UID: "4718597a-e89e-4b71-9f09-a559c94cd69d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.685730 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.685788 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.685803 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.685815 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.685826 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4718597a-e89e-4b71-9f09-a559c94cd69d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:39 crc kubenswrapper[4758]: I1203 17:18:39.685838 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzgkj\" (UniqueName: \"kubernetes.io/projected/4718597a-e89e-4b71-9f09-a559c94cd69d-kube-api-access-fzgkj\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.286987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mhwrm" event={"ID":"4718597a-e89e-4b71-9f09-a559c94cd69d","Type":"ContainerDied","Data":"3a42ea0307f61df76188ddc87516ba4ed62d1bfc7ab47d162b8503e53767a469"} Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.287051 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a42ea0307f61df76188ddc87516ba4ed62d1bfc7ab47d162b8503e53767a469" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.287072 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mhwrm" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.566169 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mhwrm"] Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.577595 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mhwrm"] Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.671040 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2rcnb"] Dec 03 17:18:40 crc kubenswrapper[4758]: E1203 17:18:40.671650 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2567e8e3-84ee-4477-94c8-c61b02086f57" containerName="init" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.671697 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2567e8e3-84ee-4477-94c8-c61b02086f57" containerName="init" Dec 03 17:18:40 crc kubenswrapper[4758]: E1203 17:18:40.671727 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" containerName="init" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.671737 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" containerName="init" Dec 03 17:18:40 crc kubenswrapper[4758]: E1203 17:18:40.671755 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4718597a-e89e-4b71-9f09-a559c94cd69d" containerName="keystone-bootstrap" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.671766 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4718597a-e89e-4b71-9f09-a559c94cd69d" containerName="keystone-bootstrap" Dec 03 17:18:40 crc kubenswrapper[4758]: E1203 17:18:40.671793 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" containerName="dnsmasq-dns" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.671805 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" containerName="dnsmasq-dns" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.672046 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2567e8e3-84ee-4477-94c8-c61b02086f57" containerName="init" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.672083 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4718597a-e89e-4b71-9f09-a559c94cd69d" containerName="keystone-bootstrap" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.672100 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a116d59-0671-4da9-8462-414abf4d5cba" containerName="dnsmasq-dns" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.672924 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.679448 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.679771 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9h99t" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.680867 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.681208 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.681251 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.698460 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2rcnb"] Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.713051 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-config-data\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.713443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-scripts\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.713508 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-combined-ca-bundle\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.713746 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-fernet-keys\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.714060 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgfmc\" (UniqueName: \"kubernetes.io/projected/fce262ba-d313-4515-ad6b-e7137a980109-kube-api-access-kgfmc\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.714157 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-credential-keys\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.814986 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-credential-keys\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.815064 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-config-data\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.815140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-scripts\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.815222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-combined-ca-bundle\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.815883 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-fernet-keys\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.815934 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgfmc\" (UniqueName: \"kubernetes.io/projected/fce262ba-d313-4515-ad6b-e7137a980109-kube-api-access-kgfmc\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.820780 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-combined-ca-bundle\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.821570 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-credential-keys\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.822691 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-fernet-keys\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.824018 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-scripts\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.833516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-config-data\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:40 crc kubenswrapper[4758]: I1203 17:18:40.840923 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgfmc\" (UniqueName: \"kubernetes.io/projected/fce262ba-d313-4515-ad6b-e7137a980109-kube-api-access-kgfmc\") pod \"keystone-bootstrap-2rcnb\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:41 crc kubenswrapper[4758]: I1203 17:18:41.005322 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:18:41 crc kubenswrapper[4758]: I1203 17:18:41.127305 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4718597a-e89e-4b71-9f09-a559c94cd69d" path="/var/lib/kubelet/pods/4718597a-e89e-4b71-9f09-a559c94cd69d/volumes" Dec 03 17:18:41 crc kubenswrapper[4758]: I1203 17:18:41.394380 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:18:41 crc kubenswrapper[4758]: I1203 17:18:41.394907 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:18:42 crc kubenswrapper[4758]: E1203 17:18:42.318605 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 03 17:18:42 crc kubenswrapper[4758]: E1203 17:18:42.318967 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gc7lk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-4jmgb_openstack(502df21c-0370-40b8-b3ef-0e55992b3008): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:18:42 crc kubenswrapper[4758]: E1203 17:18:42.320214 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-4jmgb" podUID="502df21c-0370-40b8-b3ef-0e55992b3008" Dec 03 17:18:42 crc kubenswrapper[4758]: E1203 17:18:42.812871 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 17:18:42 crc kubenswrapper[4758]: E1203 17:18:42.813149 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n677hc8h5dfh8h5b6h586h58dh7bh674hfdh5b4h5h545h5b7h658hf8h59fh675hd7h566hbh675h5dbh75h5fdh87h675h664h9ch64h58bh585q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sj57t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(994006af-c89c-42b1-b0b7-f6a247e83cab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:18:43 crc kubenswrapper[4758]: E1203 17:18:43.312387 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 17:18:43 crc kubenswrapper[4758]: E1203 17:18:43.312588 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7wzts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-q2vlf_openstack(6b98f774-a8d0-460e-a718-c119dbd1f203): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:18:43 crc kubenswrapper[4758]: E1203 17:18:43.313824 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-q2vlf" podUID="6b98f774-a8d0-460e-a718-c119dbd1f203" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.343883 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qr4sd" event={"ID":"eb23ba13-deef-46a6-a81d-a223ad56024b","Type":"ContainerDied","Data":"ba1a227cb327fda2467fe5ec0081b8b44443ef91cf706be0da6dfa7481447ca6"} Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.344310 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba1a227cb327fda2467fe5ec0081b8b44443ef91cf706be0da6dfa7481447ca6" Dec 03 17:18:43 crc kubenswrapper[4758]: E1203 17:18:43.346752 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-4jmgb" podUID="502df21c-0370-40b8-b3ef-0e55992b3008" Dec 03 17:18:43 crc kubenswrapper[4758]: E1203 17:18:43.346951 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-q2vlf" podUID="6b98f774-a8d0-460e-a718-c119dbd1f203" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.413792 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qr4sd" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.574955 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-db-sync-config-data\") pod \"eb23ba13-deef-46a6-a81d-a223ad56024b\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.575077 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcj97\" (UniqueName: \"kubernetes.io/projected/eb23ba13-deef-46a6-a81d-a223ad56024b-kube-api-access-tcj97\") pod \"eb23ba13-deef-46a6-a81d-a223ad56024b\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.575222 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-combined-ca-bundle\") pod \"eb23ba13-deef-46a6-a81d-a223ad56024b\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.575350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-config-data\") pod \"eb23ba13-deef-46a6-a81d-a223ad56024b\" (UID: \"eb23ba13-deef-46a6-a81d-a223ad56024b\") " Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.583093 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eb23ba13-deef-46a6-a81d-a223ad56024b" (UID: "eb23ba13-deef-46a6-a81d-a223ad56024b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.584181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb23ba13-deef-46a6-a81d-a223ad56024b-kube-api-access-tcj97" (OuterVolumeSpecName: "kube-api-access-tcj97") pod "eb23ba13-deef-46a6-a81d-a223ad56024b" (UID: "eb23ba13-deef-46a6-a81d-a223ad56024b"). InnerVolumeSpecName "kube-api-access-tcj97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.605013 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb23ba13-deef-46a6-a81d-a223ad56024b" (UID: "eb23ba13-deef-46a6-a81d-a223ad56024b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.646277 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-config-data" (OuterVolumeSpecName: "config-data") pod "eb23ba13-deef-46a6-a81d-a223ad56024b" (UID: "eb23ba13-deef-46a6-a81d-a223ad56024b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.678033 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.678084 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.678101 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcj97\" (UniqueName: \"kubernetes.io/projected/eb23ba13-deef-46a6-a81d-a223ad56024b-kube-api-access-tcj97\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:43 crc kubenswrapper[4758]: I1203 17:18:43.678112 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb23ba13-deef-46a6-a81d-a223ad56024b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:44 crc kubenswrapper[4758]: I1203 17:18:44.357260 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qr4sd" Dec 03 17:18:44 crc kubenswrapper[4758]: I1203 17:18:44.866314 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Dec 03 17:18:44 crc kubenswrapper[4758]: I1203 17:18:44.973115 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bxf5h"] Dec 03 17:18:44 crc kubenswrapper[4758]: E1203 17:18:44.973694 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb23ba13-deef-46a6-a81d-a223ad56024b" containerName="glance-db-sync" Dec 03 17:18:44 crc kubenswrapper[4758]: I1203 17:18:44.973718 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb23ba13-deef-46a6-a81d-a223ad56024b" containerName="glance-db-sync" Dec 03 17:18:44 crc kubenswrapper[4758]: I1203 17:18:44.973942 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb23ba13-deef-46a6-a81d-a223ad56024b" containerName="glance-db-sync" Dec 03 17:18:44 crc kubenswrapper[4758]: I1203 17:18:44.975163 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.026157 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bxf5h"] Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.142564 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-config\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.142891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.151429 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kks7s\" (UniqueName: \"kubernetes.io/projected/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-kube-api-access-kks7s\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.151596 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.151641 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.151793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.255091 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.255215 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kks7s\" (UniqueName: \"kubernetes.io/projected/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-kube-api-access-kks7s\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.255251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.255268 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.255308 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.255349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-config\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.256458 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-config\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.257313 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.259414 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.260487 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.260891 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.299857 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kks7s\" (UniqueName: \"kubernetes.io/projected/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-kube-api-access-kks7s\") pod \"dnsmasq-dns-56df8fb6b7-bxf5h\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.329041 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.979464 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.983275 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.993336 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.993730 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2c867" Dec 03 17:18:45 crc kubenswrapper[4758]: I1203 17:18:45.993336 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.011337 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.144533 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.146322 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.152770 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.163970 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.184936 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-logs\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.185008 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.185125 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.185223 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whclj\" (UniqueName: \"kubernetes.io/projected/147df672-7ed8-41bd-b083-ca3ef2752ce0-kube-api-access-whclj\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.185276 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.185316 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-config-data\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.185362 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-scripts\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287070 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287135 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287164 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-config-data\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287186 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-scripts\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287242 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-logs\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287282 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-logs\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287301 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287376 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287408 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287447 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287474 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287493 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r494\" (UniqueName: \"kubernetes.io/projected/b18488c2-1909-4155-8bec-89b3556a8c14-kube-api-access-4r494\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287525 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whclj\" (UniqueName: \"kubernetes.io/projected/147df672-7ed8-41bd-b083-ca3ef2752ce0-kube-api-access-whclj\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.287879 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.288491 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.289530 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-logs\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.298588 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.303403 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-config-data\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.303748 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-scripts\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.312390 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whclj\" (UniqueName: \"kubernetes.io/projected/147df672-7ed8-41bd-b083-ca3ef2752ce0-kube-api-access-whclj\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.332266 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " pod="openstack/glance-default-external-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.389139 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-logs\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.389363 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.389391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.389772 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-logs\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.390011 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.390343 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.390376 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r494\" (UniqueName: \"kubernetes.io/projected/b18488c2-1909-4155-8bec-89b3556a8c14-kube-api-access-4r494\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.390481 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.390512 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.391085 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.396209 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.401171 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.416524 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.423516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r494\" (UniqueName: \"kubernetes.io/projected/b18488c2-1909-4155-8bec-89b3556a8c14-kube-api-access-4r494\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.445946 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.486310 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:18:46 crc kubenswrapper[4758]: I1203 17:18:46.622124 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:18:47 crc kubenswrapper[4758]: I1203 17:18:47.808432 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:18:47 crc kubenswrapper[4758]: I1203 17:18:47.959573 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:18:49 crc kubenswrapper[4758]: I1203 17:18:49.867085 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Dec 03 17:18:54 crc kubenswrapper[4758]: I1203 17:18:54.868332 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Dec 03 17:18:54 crc kubenswrapper[4758]: I1203 17:18:54.869518 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.005323 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.078255 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-nb\") pod \"566c8d53-04a9-4937-bdab-4cac68a6ca49\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.078340 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2twf4\" (UniqueName: \"kubernetes.io/projected/566c8d53-04a9-4937-bdab-4cac68a6ca49-kube-api-access-2twf4\") pod \"566c8d53-04a9-4937-bdab-4cac68a6ca49\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.078375 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-config\") pod \"566c8d53-04a9-4937-bdab-4cac68a6ca49\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.078468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-sb\") pod \"566c8d53-04a9-4937-bdab-4cac68a6ca49\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.078538 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-dns-svc\") pod \"566c8d53-04a9-4937-bdab-4cac68a6ca49\" (UID: \"566c8d53-04a9-4937-bdab-4cac68a6ca49\") " Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.090073 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566c8d53-04a9-4937-bdab-4cac68a6ca49-kube-api-access-2twf4" (OuterVolumeSpecName: "kube-api-access-2twf4") pod "566c8d53-04a9-4937-bdab-4cac68a6ca49" (UID: "566c8d53-04a9-4937-bdab-4cac68a6ca49"). InnerVolumeSpecName "kube-api-access-2twf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.141996 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "566c8d53-04a9-4937-bdab-4cac68a6ca49" (UID: "566c8d53-04a9-4937-bdab-4cac68a6ca49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.142111 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "566c8d53-04a9-4937-bdab-4cac68a6ca49" (UID: "566c8d53-04a9-4937-bdab-4cac68a6ca49"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.142111 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "566c8d53-04a9-4937-bdab-4cac68a6ca49" (UID: "566c8d53-04a9-4937-bdab-4cac68a6ca49"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.165421 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-config" (OuterVolumeSpecName: "config") pod "566c8d53-04a9-4937-bdab-4cac68a6ca49" (UID: "566c8d53-04a9-4937-bdab-4cac68a6ca49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.181442 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.181659 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2twf4\" (UniqueName: \"kubernetes.io/projected/566c8d53-04a9-4937-bdab-4cac68a6ca49-kube-api-access-2twf4\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.181769 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.181849 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.181976 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566c8d53-04a9-4937-bdab-4cac68a6ca49-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.499145 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" event={"ID":"566c8d53-04a9-4937-bdab-4cac68a6ca49","Type":"ContainerDied","Data":"1cf0016e1f0b774f19cba679fc7b3cb41009855c5aea535c21e98ca34e4d9bf9"} Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.499199 4758 scope.go:117] "RemoveContainer" containerID="e2b441905ca3a543bb1a2a712ed67ede3cc85561f57b550f7e41f589e4531e8a" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.499279 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.552532 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-zpczk"] Dec 03 17:18:55 crc kubenswrapper[4758]: I1203 17:18:55.564811 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-zpczk"] Dec 03 17:18:57 crc kubenswrapper[4758]: I1203 17:18:57.128798 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" path="/var/lib/kubelet/pods/566c8d53-04a9-4937-bdab-4cac68a6ca49/volumes" Dec 03 17:18:58 crc kubenswrapper[4758]: E1203 17:18:58.470559 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 17:18:58 crc kubenswrapper[4758]: E1203 17:18:58.471349 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2cl5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-6l8jp_openstack(dbfaa3a1-7e61-4537-b953-07269f386adc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:18:58 crc kubenswrapper[4758]: E1203 17:18:58.472736 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-6l8jp" podUID="dbfaa3a1-7e61-4537-b953-07269f386adc" Dec 03 17:18:58 crc kubenswrapper[4758]: E1203 17:18:58.558841 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-6l8jp" podUID="dbfaa3a1-7e61-4537-b953-07269f386adc" Dec 03 17:18:58 crc kubenswrapper[4758]: I1203 17:18:58.559337 4758 scope.go:117] "RemoveContainer" containerID="50338516a62f4c77d97174e64b721bc57c9c6d94dadca1ad11e722975542178f" Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.031184 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2rcnb"] Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.157352 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.546443 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2rcnb" event={"ID":"fce262ba-d313-4515-ad6b-e7137a980109","Type":"ContainerStarted","Data":"cc234173c0c1f85c77be92341554d1a75f3b1857783d167b5ab17235b9336237"} Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.547001 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2rcnb" event={"ID":"fce262ba-d313-4515-ad6b-e7137a980109","Type":"ContainerStarted","Data":"40187032a8a62b4287e04e4067536a320353e13caa0c8ac5de9a32bfbbef4ec7"} Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.552510 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4jmgb" event={"ID":"502df21c-0370-40b8-b3ef-0e55992b3008","Type":"ContainerStarted","Data":"986d8b4d0b76748ce94c47a15f1571bda7e27fd288e0506f7528909dffe76a1b"} Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.574263 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2rcnb" podStartSLOduration=19.574226677 podStartE2EDuration="19.574226677s" podCreationTimestamp="2025-12-03 17:18:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:59.570396543 +0000 UTC m=+1394.771773404" watchObservedRunningTime="2025-12-03 17:18:59.574226677 +0000 UTC m=+1394.775603548" Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.638656 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4jmgb" podStartSLOduration=2.826876464 podStartE2EDuration="33.638615437s" podCreationTimestamp="2025-12-03 17:18:26 +0000 UTC" firstStartedPulling="2025-12-03 17:18:28.349687076 +0000 UTC m=+1363.551063927" lastFinishedPulling="2025-12-03 17:18:59.161426039 +0000 UTC m=+1394.362802900" observedRunningTime="2025-12-03 17:18:59.593598611 +0000 UTC m=+1394.794975472" watchObservedRunningTime="2025-12-03 17:18:59.638615437 +0000 UTC m=+1394.839992298" Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.650156 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bxf5h"] Dec 03 17:18:59 crc kubenswrapper[4758]: W1203 17:18:59.716349 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4401ac8b_420e_41b9_a6b0_aba2886b8ff8.slice/crio-cc1b52602f084ab721084d07659a360a9a4947eadf99b3c98f294744a9a1a01d WatchSource:0}: Error finding container cc1b52602f084ab721084d07659a360a9a4947eadf99b3c98f294744a9a1a01d: Status 404 returned error can't find the container with id cc1b52602f084ab721084d07659a360a9a4947eadf99b3c98f294744a9a1a01d Dec 03 17:18:59 crc kubenswrapper[4758]: W1203 17:18:59.793062 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod147df672_7ed8_41bd_b083_ca3ef2752ce0.slice/crio-d41b01f3a2df786278dd2d08db09cd891bbfec4b8f545995259e6758689da75d WatchSource:0}: Error finding container d41b01f3a2df786278dd2d08db09cd891bbfec4b8f545995259e6758689da75d: Status 404 returned error can't find the container with id d41b01f3a2df786278dd2d08db09cd891bbfec4b8f545995259e6758689da75d Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.819421 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:18:59 crc kubenswrapper[4758]: I1203 17:18:59.869578 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-zpczk" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.573356 4758 generic.go:334] "Generic (PLEG): container finished" podID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerID="50b6eae62c328e1085de8a54db7c2d8a55349a548197be9d1d11098b72f55ec6" exitCode=0 Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.573817 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" event={"ID":"4401ac8b-420e-41b9-a6b0-aba2886b8ff8","Type":"ContainerDied","Data":"50b6eae62c328e1085de8a54db7c2d8a55349a548197be9d1d11098b72f55ec6"} Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.573867 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" event={"ID":"4401ac8b-420e-41b9-a6b0-aba2886b8ff8","Type":"ContainerStarted","Data":"cc1b52602f084ab721084d07659a360a9a4947eadf99b3c98f294744a9a1a01d"} Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.580361 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-q2vlf" event={"ID":"6b98f774-a8d0-460e-a718-c119dbd1f203","Type":"ContainerStarted","Data":"c4f33dad18cd24c6fb9262563f5df5def2e2efa7c561f64389ce9ac18f00ff1c"} Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.592858 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerStarted","Data":"e28be1b4932b60511077a9023fc129e0953206f967f40a629526334ce7f94d00"} Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.598247 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"147df672-7ed8-41bd-b083-ca3ef2752ce0","Type":"ContainerStarted","Data":"d41b01f3a2df786278dd2d08db09cd891bbfec4b8f545995259e6758689da75d"} Dec 03 17:19:00 crc kubenswrapper[4758]: W1203 17:19:00.630972 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb18488c2_1909_4155_8bec_89b3556a8c14.slice/crio-126c99cc4f0ff4603949e318147906394145fb5151ef304e8a78c4da626590a0 WatchSource:0}: Error finding container 126c99cc4f0ff4603949e318147906394145fb5151ef304e8a78c4da626590a0: Status 404 returned error can't find the container with id 126c99cc4f0ff4603949e318147906394145fb5151ef304e8a78c4da626590a0 Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.678321 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:19:00 crc kubenswrapper[4758]: I1203 17:19:00.693834 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-q2vlf" podStartSLOduration=3.205418296 podStartE2EDuration="34.693806568s" podCreationTimestamp="2025-12-03 17:18:26 +0000 UTC" firstStartedPulling="2025-12-03 17:18:28.301996287 +0000 UTC m=+1363.503373148" lastFinishedPulling="2025-12-03 17:18:59.790384559 +0000 UTC m=+1394.991761420" observedRunningTime="2025-12-03 17:19:00.648488613 +0000 UTC m=+1395.849865474" watchObservedRunningTime="2025-12-03 17:19:00.693806568 +0000 UTC m=+1395.895183429" Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.626856 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"147df672-7ed8-41bd-b083-ca3ef2752ce0","Type":"ContainerStarted","Data":"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc"} Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.627719 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"147df672-7ed8-41bd-b083-ca3ef2752ce0","Type":"ContainerStarted","Data":"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a"} Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.627878 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-log" containerID="cri-o://c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a" gracePeriod=30 Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.628446 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-httpd" containerID="cri-o://39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc" gracePeriod=30 Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.635329 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" event={"ID":"4401ac8b-420e-41b9-a6b0-aba2886b8ff8","Type":"ContainerStarted","Data":"25065c61a15ddba760fbd565f929949c6cd3b972d95f83ea74bea465a8ba22dc"} Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.636259 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.640545 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b18488c2-1909-4155-8bec-89b3556a8c14","Type":"ContainerStarted","Data":"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b"} Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.640592 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b18488c2-1909-4155-8bec-89b3556a8c14","Type":"ContainerStarted","Data":"126c99cc4f0ff4603949e318147906394145fb5151ef304e8a78c4da626590a0"} Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.656184 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.65616108 podStartE2EDuration="17.65616108s" podCreationTimestamp="2025-12-03 17:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:01.65023192 +0000 UTC m=+1396.851608801" watchObservedRunningTime="2025-12-03 17:19:01.65616108 +0000 UTC m=+1396.857537951" Dec 03 17:19:01 crc kubenswrapper[4758]: I1203 17:19:01.691973 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" podStartSLOduration=17.691950637 podStartE2EDuration="17.691950637s" podCreationTimestamp="2025-12-03 17:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:01.680767455 +0000 UTC m=+1396.882144316" watchObservedRunningTime="2025-12-03 17:19:01.691950637 +0000 UTC m=+1396.893327498" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.365104 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.452572 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.452663 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-httpd-run\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.452775 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-config-data\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.452808 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-scripts\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.453014 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-combined-ca-bundle\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.453070 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-logs\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.453121 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whclj\" (UniqueName: \"kubernetes.io/projected/147df672-7ed8-41bd-b083-ca3ef2752ce0-kube-api-access-whclj\") pod \"147df672-7ed8-41bd-b083-ca3ef2752ce0\" (UID: \"147df672-7ed8-41bd-b083-ca3ef2752ce0\") " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.456540 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-logs" (OuterVolumeSpecName: "logs") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.461555 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.469789 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-scripts" (OuterVolumeSpecName: "scripts") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.477279 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.480315 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147df672-7ed8-41bd-b083-ca3ef2752ce0-kube-api-access-whclj" (OuterVolumeSpecName: "kube-api-access-whclj") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "kube-api-access-whclj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.549131 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.556324 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.556639 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.556748 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.556816 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whclj\" (UniqueName: \"kubernetes.io/projected/147df672-7ed8-41bd-b083-ca3ef2752ce0-kube-api-access-whclj\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.556890 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.556950 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/147df672-7ed8-41bd-b083-ca3ef2752ce0-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.586646 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.594156 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-config-data" (OuterVolumeSpecName: "config-data") pod "147df672-7ed8-41bd-b083-ca3ef2752ce0" (UID: "147df672-7ed8-41bd-b083-ca3ef2752ce0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.658707 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.658742 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/147df672-7ed8-41bd-b083-ca3ef2752ce0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.660055 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b18488c2-1909-4155-8bec-89b3556a8c14","Type":"ContainerStarted","Data":"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7"} Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.660310 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-log" containerID="cri-o://524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b" gracePeriod=30 Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.660462 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-httpd" containerID="cri-o://7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7" gracePeriod=30 Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.674550 4758 generic.go:334] "Generic (PLEG): container finished" podID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerID="39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc" exitCode=143 Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.674591 4758 generic.go:334] "Generic (PLEG): container finished" podID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerID="c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a" exitCode=143 Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.675879 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.675915 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"147df672-7ed8-41bd-b083-ca3ef2752ce0","Type":"ContainerDied","Data":"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc"} Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.675957 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"147df672-7ed8-41bd-b083-ca3ef2752ce0","Type":"ContainerDied","Data":"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a"} Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.675970 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"147df672-7ed8-41bd-b083-ca3ef2752ce0","Type":"ContainerDied","Data":"d41b01f3a2df786278dd2d08db09cd891bbfec4b8f545995259e6758689da75d"} Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.675991 4758 scope.go:117] "RemoveContainer" containerID="39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.754373 4758 scope.go:117] "RemoveContainer" containerID="c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.806511 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.806479092 podStartE2EDuration="17.806479092s" podCreationTimestamp="2025-12-03 17:18:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:02.691090453 +0000 UTC m=+1397.892467334" watchObservedRunningTime="2025-12-03 17:19:02.806479092 +0000 UTC m=+1398.007855953" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.820652 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.829171 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.840511 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:19:02 crc kubenswrapper[4758]: E1203 17:19:02.841107 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-httpd" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841135 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-httpd" Dec 03 17:19:02 crc kubenswrapper[4758]: E1203 17:19:02.841154 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="init" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841162 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="init" Dec 03 17:19:02 crc kubenswrapper[4758]: E1203 17:19:02.841179 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841185 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" Dec 03 17:19:02 crc kubenswrapper[4758]: E1203 17:19:02.841195 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-log" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841203 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-log" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841435 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-log" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841464 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="566c8d53-04a9-4937-bdab-4cac68a6ca49" containerName="dnsmasq-dns" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.841616 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" containerName="glance-httpd" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.843081 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.846424 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.846436 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.861213 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.880631 4758 scope.go:117] "RemoveContainer" containerID="39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc" Dec 03 17:19:02 crc kubenswrapper[4758]: E1203 17:19:02.881311 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc\": container with ID starting with 39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc not found: ID does not exist" containerID="39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.881378 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc"} err="failed to get container status \"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc\": rpc error: code = NotFound desc = could not find container \"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc\": container with ID starting with 39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc not found: ID does not exist" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.881411 4758 scope.go:117] "RemoveContainer" containerID="c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a" Dec 03 17:19:02 crc kubenswrapper[4758]: E1203 17:19:02.881696 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a\": container with ID starting with c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a not found: ID does not exist" containerID="c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.881723 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a"} err="failed to get container status \"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a\": rpc error: code = NotFound desc = could not find container \"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a\": container with ID starting with c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a not found: ID does not exist" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.881738 4758 scope.go:117] "RemoveContainer" containerID="39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.883211 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc"} err="failed to get container status \"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc\": rpc error: code = NotFound desc = could not find container \"39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc\": container with ID starting with 39065391d21026e118bd857fc27d63ee287ac70c867d7fa1e2aa3a765f8a1edc not found: ID does not exist" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.883242 4758 scope.go:117] "RemoveContainer" containerID="c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.883537 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a"} err="failed to get container status \"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a\": rpc error: code = NotFound desc = could not find container \"c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a\": container with ID starting with c3e30765c7b4fd4b8421290f4f69ab3c0157a6a37022c632843a637cc8c51c1a not found: ID does not exist" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975308 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975438 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8bf5\" (UniqueName: \"kubernetes.io/projected/74ca64c8-28c7-472e-85bd-acd6af307580-kube-api-access-n8bf5\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975483 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-config-data\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975510 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-scripts\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975624 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975657 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975821 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-logs\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:02 crc kubenswrapper[4758]: I1203 17:19:02.975849 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078501 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078563 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078631 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-logs\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078720 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078770 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8bf5\" (UniqueName: \"kubernetes.io/projected/74ca64c8-28c7-472e-85bd-acd6af307580-kube-api-access-n8bf5\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078799 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-config-data\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.078836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-scripts\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.084614 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.085046 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.085481 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-logs\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.097745 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-scripts\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.098380 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.101516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.108572 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8bf5\" (UniqueName: \"kubernetes.io/projected/74ca64c8-28c7-472e-85bd-acd6af307580-kube-api-access-n8bf5\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.110983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-config-data\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.136329 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.156974 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147df672-7ed8-41bd-b083-ca3ef2752ce0" path="/var/lib/kubelet/pods/147df672-7ed8-41bd-b083-ca3ef2752ce0/volumes" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.182661 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.283128 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.388702 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-combined-ca-bundle\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.388858 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-config-data\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.388926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-logs\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.388971 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-scripts\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.389006 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-httpd-run\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.389051 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.389205 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r494\" (UniqueName: \"kubernetes.io/projected/b18488c2-1909-4155-8bec-89b3556a8c14-kube-api-access-4r494\") pod \"b18488c2-1909-4155-8bec-89b3556a8c14\" (UID: \"b18488c2-1909-4155-8bec-89b3556a8c14\") " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.392580 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.392924 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-logs" (OuterVolumeSpecName: "logs") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.395185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18488c2-1909-4155-8bec-89b3556a8c14-kube-api-access-4r494" (OuterVolumeSpecName: "kube-api-access-4r494") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "kube-api-access-4r494". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.406548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.421035 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-scripts" (OuterVolumeSpecName: "scripts") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.474297 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.492630 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r494\" (UniqueName: \"kubernetes.io/projected/b18488c2-1909-4155-8bec-89b3556a8c14-kube-api-access-4r494\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.492785 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.492805 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.492817 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.492829 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b18488c2-1909-4155-8bec-89b3556a8c14-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.492879 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.503548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-config-data" (OuterVolumeSpecName: "config-data") pod "b18488c2-1909-4155-8bec-89b3556a8c14" (UID: "b18488c2-1909-4155-8bec-89b3556a8c14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.522803 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.596888 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b18488c2-1909-4155-8bec-89b3556a8c14-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.596939 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704547 4758 generic.go:334] "Generic (PLEG): container finished" podID="b18488c2-1909-4155-8bec-89b3556a8c14" containerID="7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7" exitCode=143 Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704599 4758 generic.go:334] "Generic (PLEG): container finished" podID="b18488c2-1909-4155-8bec-89b3556a8c14" containerID="524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b" exitCode=143 Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704652 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704652 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b18488c2-1909-4155-8bec-89b3556a8c14","Type":"ContainerDied","Data":"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7"} Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704810 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b18488c2-1909-4155-8bec-89b3556a8c14","Type":"ContainerDied","Data":"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b"} Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704823 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b18488c2-1909-4155-8bec-89b3556a8c14","Type":"ContainerDied","Data":"126c99cc4f0ff4603949e318147906394145fb5151ef304e8a78c4da626590a0"} Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.704850 4758 scope.go:117] "RemoveContainer" containerID="7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.756158 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.758255 4758 scope.go:117] "RemoveContainer" containerID="524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.794924 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.831344 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:19:03 crc kubenswrapper[4758]: E1203 17:19:03.831958 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-httpd" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.831985 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-httpd" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.834379 4758 scope.go:117] "RemoveContainer" containerID="7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7" Dec 03 17:19:03 crc kubenswrapper[4758]: E1203 17:19:03.834623 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-log" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.834639 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-log" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.834909 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-log" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.834929 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" containerName="glance-httpd" Dec 03 17:19:03 crc kubenswrapper[4758]: E1203 17:19:03.844892 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7\": container with ID starting with 7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7 not found: ID does not exist" containerID="7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.844945 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7"} err="failed to get container status \"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7\": rpc error: code = NotFound desc = could not find container \"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7\": container with ID starting with 7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7 not found: ID does not exist" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.844974 4758 scope.go:117] "RemoveContainer" containerID="524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.845956 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:03 crc kubenswrapper[4758]: E1203 17:19:03.847238 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b\": container with ID starting with 524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b not found: ID does not exist" containerID="524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.847273 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b"} err="failed to get container status \"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b\": rpc error: code = NotFound desc = could not find container \"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b\": container with ID starting with 524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b not found: ID does not exist" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.847300 4758 scope.go:117] "RemoveContainer" containerID="7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.847771 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7"} err="failed to get container status \"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7\": rpc error: code = NotFound desc = could not find container \"7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7\": container with ID starting with 7618790b90e38eeb50f506b3476db3cf3b1bcf8332bad4fa294dbf55e30ecda7 not found: ID does not exist" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.847789 4758 scope.go:117] "RemoveContainer" containerID="524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.848174 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b"} err="failed to get container status \"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b\": rpc error: code = NotFound desc = could not find container \"524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b\": container with ID starting with 524b7cebcdd27fd5729c126908cf4e8da7c445ef487bea0cc5a66b0cb6c23a9b not found: ID does not exist" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.850267 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.855808 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.882623 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:19:03 crc kubenswrapper[4758]: I1203 17:19:03.949938 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:19:03 crc kubenswrapper[4758]: W1203 17:19:03.953373 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74ca64c8_28c7_472e_85bd_acd6af307580.slice/crio-89b58a21b26fe2a2527553d5556f7d694876e49729ca2f5d4fe36850040209d1 WatchSource:0}: Error finding container 89b58a21b26fe2a2527553d5556f7d694876e49729ca2f5d4fe36850040209d1: Status 404 returned error can't find the container with id 89b58a21b26fe2a2527553d5556f7d694876e49729ca2f5d4fe36850040209d1 Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.006451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.006974 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.007064 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.007106 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-logs\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.007242 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.007267 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tllxp\" (UniqueName: \"kubernetes.io/projected/12d6c948-8253-4dbf-b782-27603b7ae061-kube-api-access-tllxp\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.007301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-config-data\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.007369 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-scripts\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109579 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109657 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-logs\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109723 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109744 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tllxp\" (UniqueName: \"kubernetes.io/projected/12d6c948-8253-4dbf-b782-27603b7ae061-kube-api-access-tllxp\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109773 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-config-data\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109806 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-scripts\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109875 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.109924 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.110425 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.110855 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-logs\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.111123 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.120139 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.120444 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.125464 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-config-data\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.130910 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-scripts\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.139146 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tllxp\" (UniqueName: \"kubernetes.io/projected/12d6c948-8253-4dbf-b782-27603b7ae061-kube-api-access-tllxp\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.165629 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.183033 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.738159 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ca64c8-28c7-472e-85bd-acd6af307580","Type":"ContainerStarted","Data":"89b58a21b26fe2a2527553d5556f7d694876e49729ca2f5d4fe36850040209d1"} Dec 03 17:19:04 crc kubenswrapper[4758]: I1203 17:19:04.974984 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.139533 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18488c2-1909-4155-8bec-89b3556a8c14" path="/var/lib/kubelet/pods/b18488c2-1909-4155-8bec-89b3556a8c14/volumes" Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.349902 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.432714 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4ggvx"] Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.433114 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="dnsmasq-dns" containerID="cri-o://753c1f65ee759c0fdb55ace0c17c7d87b84db76d9c268e39c820bbac1853d21b" gracePeriod=10 Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.761015 4758 generic.go:334] "Generic (PLEG): container finished" podID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerID="753c1f65ee759c0fdb55ace0c17c7d87b84db76d9c268e39c820bbac1853d21b" exitCode=0 Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.761148 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" event={"ID":"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9","Type":"ContainerDied","Data":"753c1f65ee759c0fdb55ace0c17c7d87b84db76d9c268e39c820bbac1853d21b"} Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.766831 4758 generic.go:334] "Generic (PLEG): container finished" podID="fce262ba-d313-4515-ad6b-e7137a980109" containerID="cc234173c0c1f85c77be92341554d1a75f3b1857783d167b5ab17235b9336237" exitCode=0 Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.766964 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2rcnb" event={"ID":"fce262ba-d313-4515-ad6b-e7137a980109","Type":"ContainerDied","Data":"cc234173c0c1f85c77be92341554d1a75f3b1857783d167b5ab17235b9336237"} Dec 03 17:19:05 crc kubenswrapper[4758]: I1203 17:19:05.771374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ca64c8-28c7-472e-85bd-acd6af307580","Type":"ContainerStarted","Data":"55c59d419d0dddb9e20a6a9054a34c3b529a7c39deb72363a316ff284231e1fe"} Dec 03 17:19:07 crc kubenswrapper[4758]: I1203 17:19:07.578735 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.144:5353: connect: connection refused" Dec 03 17:19:09 crc kubenswrapper[4758]: I1203 17:19:09.825621 4758 generic.go:334] "Generic (PLEG): container finished" podID="502df21c-0370-40b8-b3ef-0e55992b3008" containerID="986d8b4d0b76748ce94c47a15f1571bda7e27fd288e0506f7528909dffe76a1b" exitCode=0 Dec 03 17:19:09 crc kubenswrapper[4758]: I1203 17:19:09.826064 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4jmgb" event={"ID":"502df21c-0370-40b8-b3ef-0e55992b3008","Type":"ContainerDied","Data":"986d8b4d0b76748ce94c47a15f1571bda7e27fd288e0506f7528909dffe76a1b"} Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.394699 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.395191 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.395243 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.396307 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"edeb1f97ccdd747653b82f053447c5a4a38ed821ec3f1cb524dc9338c014ae57"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.396384 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://edeb1f97ccdd747653b82f053447c5a4a38ed821ec3f1cb524dc9338c014ae57" gracePeriod=600 Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.604567 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.609635 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4jmgb" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.722877 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-config-data\") pod \"502df21c-0370-40b8-b3ef-0e55992b3008\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.722965 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-credential-keys\") pod \"fce262ba-d313-4515-ad6b-e7137a980109\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723011 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-scripts\") pod \"502df21c-0370-40b8-b3ef-0e55992b3008\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723083 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/502df21c-0370-40b8-b3ef-0e55992b3008-logs\") pod \"502df21c-0370-40b8-b3ef-0e55992b3008\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723152 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-combined-ca-bundle\") pod \"502df21c-0370-40b8-b3ef-0e55992b3008\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723180 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-combined-ca-bundle\") pod \"fce262ba-d313-4515-ad6b-e7137a980109\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723219 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-scripts\") pod \"fce262ba-d313-4515-ad6b-e7137a980109\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723288 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgfmc\" (UniqueName: \"kubernetes.io/projected/fce262ba-d313-4515-ad6b-e7137a980109-kube-api-access-kgfmc\") pod \"fce262ba-d313-4515-ad6b-e7137a980109\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723409 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-config-data\") pod \"fce262ba-d313-4515-ad6b-e7137a980109\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723451 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-fernet-keys\") pod \"fce262ba-d313-4515-ad6b-e7137a980109\" (UID: \"fce262ba-d313-4515-ad6b-e7137a980109\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.723536 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc7lk\" (UniqueName: \"kubernetes.io/projected/502df21c-0370-40b8-b3ef-0e55992b3008-kube-api-access-gc7lk\") pod \"502df21c-0370-40b8-b3ef-0e55992b3008\" (UID: \"502df21c-0370-40b8-b3ef-0e55992b3008\") " Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.724567 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/502df21c-0370-40b8-b3ef-0e55992b3008-logs" (OuterVolumeSpecName: "logs") pod "502df21c-0370-40b8-b3ef-0e55992b3008" (UID: "502df21c-0370-40b8-b3ef-0e55992b3008"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.727343 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/502df21c-0370-40b8-b3ef-0e55992b3008-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.733436 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fce262ba-d313-4515-ad6b-e7137a980109-kube-api-access-kgfmc" (OuterVolumeSpecName: "kube-api-access-kgfmc") pod "fce262ba-d313-4515-ad6b-e7137a980109" (UID: "fce262ba-d313-4515-ad6b-e7137a980109"). InnerVolumeSpecName "kube-api-access-kgfmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.733575 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502df21c-0370-40b8-b3ef-0e55992b3008-kube-api-access-gc7lk" (OuterVolumeSpecName: "kube-api-access-gc7lk") pod "502df21c-0370-40b8-b3ef-0e55992b3008" (UID: "502df21c-0370-40b8-b3ef-0e55992b3008"). InnerVolumeSpecName "kube-api-access-gc7lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.793608 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fce262ba-d313-4515-ad6b-e7137a980109" (UID: "fce262ba-d313-4515-ad6b-e7137a980109"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.793844 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-scripts" (OuterVolumeSpecName: "scripts") pod "fce262ba-d313-4515-ad6b-e7137a980109" (UID: "fce262ba-d313-4515-ad6b-e7137a980109"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.796603 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-scripts" (OuterVolumeSpecName: "scripts") pod "502df21c-0370-40b8-b3ef-0e55992b3008" (UID: "502df21c-0370-40b8-b3ef-0e55992b3008"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.798949 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-config-data" (OuterVolumeSpecName: "config-data") pod "fce262ba-d313-4515-ad6b-e7137a980109" (UID: "fce262ba-d313-4515-ad6b-e7137a980109"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.800627 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fce262ba-d313-4515-ad6b-e7137a980109" (UID: "fce262ba-d313-4515-ad6b-e7137a980109"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.805271 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fce262ba-d313-4515-ad6b-e7137a980109" (UID: "fce262ba-d313-4515-ad6b-e7137a980109"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.810773 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "502df21c-0370-40b8-b3ef-0e55992b3008" (UID: "502df21c-0370-40b8-b3ef-0e55992b3008"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.828121 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-config-data" (OuterVolumeSpecName: "config-data") pod "502df21c-0370-40b8-b3ef-0e55992b3008" (UID: "502df21c-0370-40b8-b3ef-0e55992b3008"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829616 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829708 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829722 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc7lk\" (UniqueName: \"kubernetes.io/projected/502df21c-0370-40b8-b3ef-0e55992b3008-kube-api-access-gc7lk\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829734 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829744 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829753 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829769 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502df21c-0370-40b8-b3ef-0e55992b3008-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829778 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829786 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce262ba-d313-4515-ad6b-e7137a980109-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.829794 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgfmc\" (UniqueName: \"kubernetes.io/projected/fce262ba-d313-4515-ad6b-e7137a980109-kube-api-access-kgfmc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.859449 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"12d6c948-8253-4dbf-b782-27603b7ae061","Type":"ContainerStarted","Data":"e6d6992da89cfbc258334346183f4d711c90d32212a8532b19cbfb896282e30e"} Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.864079 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" event={"ID":"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9","Type":"ContainerDied","Data":"ffec55692f70664f94aed85a16b45310194554b9829fa8e481a713dfdfa9d505"} Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.864126 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffec55692f70664f94aed85a16b45310194554b9829fa8e481a713dfdfa9d505" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.875107 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2rcnb" event={"ID":"fce262ba-d313-4515-ad6b-e7137a980109","Type":"ContainerDied","Data":"40187032a8a62b4287e04e4067536a320353e13caa0c8ac5de9a32bfbbef4ec7"} Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.875167 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40187032a8a62b4287e04e4067536a320353e13caa0c8ac5de9a32bfbbef4ec7" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.875191 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2rcnb" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.901910 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="edeb1f97ccdd747653b82f053447c5a4a38ed821ec3f1cb524dc9338c014ae57" exitCode=0 Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.901982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"edeb1f97ccdd747653b82f053447c5a4a38ed821ec3f1cb524dc9338c014ae57"} Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.902550 4758 scope.go:117] "RemoveContainer" containerID="74575219fc052d90d1263cdeb0cdc0186ad14da295a8ec4c660422997cca91c3" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.904160 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.911888 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4jmgb" event={"ID":"502df21c-0370-40b8-b3ef-0e55992b3008","Type":"ContainerDied","Data":"550829c342617ab5ac095f4de671c710c04dbc81e2d2657d218df85ed1fa8f5f"} Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.911953 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="550829c342617ab5ac095f4de671c710c04dbc81e2d2657d218df85ed1fa8f5f" Dec 03 17:19:11 crc kubenswrapper[4758]: I1203 17:19:11.912046 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4jmgb" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.018457 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68d78d4d4d-rhzqc"] Dec 03 17:19:12 crc kubenswrapper[4758]: E1203 17:19:12.018962 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce262ba-d313-4515-ad6b-e7137a980109" containerName="keystone-bootstrap" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.018984 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce262ba-d313-4515-ad6b-e7137a980109" containerName="keystone-bootstrap" Dec 03 17:19:12 crc kubenswrapper[4758]: E1203 17:19:12.019007 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="dnsmasq-dns" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.019013 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="dnsmasq-dns" Dec 03 17:19:12 crc kubenswrapper[4758]: E1203 17:19:12.019723 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="init" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.019736 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="init" Dec 03 17:19:12 crc kubenswrapper[4758]: E1203 17:19:12.019756 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502df21c-0370-40b8-b3ef-0e55992b3008" containerName="placement-db-sync" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.019785 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="502df21c-0370-40b8-b3ef-0e55992b3008" containerName="placement-db-sync" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.020113 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce262ba-d313-4515-ad6b-e7137a980109" containerName="keystone-bootstrap" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.020146 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="502df21c-0370-40b8-b3ef-0e55992b3008" containerName="placement-db-sync" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.020183 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" containerName="dnsmasq-dns" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.022382 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.027295 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.027324 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.029133 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.029147 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.030964 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-87xxm" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.035910 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-swift-storage-0\") pod \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.036412 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-nb\") pod \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.036481 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4czhn\" (UniqueName: \"kubernetes.io/projected/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-kube-api-access-4czhn\") pod \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.036586 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-config\") pod \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.036870 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-svc\") pod \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.039929 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68d78d4d4d-rhzqc"] Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.036916 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-sb\") pod \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\" (UID: \"9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9\") " Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.114828 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-kube-api-access-4czhn" (OuterVolumeSpecName: "kube-api-access-4czhn") pod "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" (UID: "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9"). InnerVolumeSpecName "kube-api-access-4czhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.148858 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-public-tls-certs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.149153 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-scripts\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.149793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srjfp\" (UniqueName: \"kubernetes.io/projected/1f328daf-ac93-432b-8645-14b7030a24d1-kube-api-access-srjfp\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.150044 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-config-data\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.150301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f328daf-ac93-432b-8645-14b7030a24d1-logs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.150340 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-combined-ca-bundle\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.150366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-internal-tls-certs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.150461 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4czhn\" (UniqueName: \"kubernetes.io/projected/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-kube-api-access-4czhn\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.252721 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-scripts\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.252888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srjfp\" (UniqueName: \"kubernetes.io/projected/1f328daf-ac93-432b-8645-14b7030a24d1-kube-api-access-srjfp\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.253086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-config-data\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.253320 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f328daf-ac93-432b-8645-14b7030a24d1-logs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.253384 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-combined-ca-bundle\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.253416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-internal-tls-certs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.253529 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-public-tls-certs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.254270 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f328daf-ac93-432b-8645-14b7030a24d1-logs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.264493 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-scripts\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.269277 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-internal-tls-certs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.271600 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-config-data\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.275604 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-combined-ca-bundle\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.280708 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-public-tls-certs\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.281198 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srjfp\" (UniqueName: \"kubernetes.io/projected/1f328daf-ac93-432b-8645-14b7030a24d1-kube-api-access-srjfp\") pod \"placement-68d78d4d4d-rhzqc\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.298817 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" (UID: "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.310207 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" (UID: "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.311885 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" (UID: "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.314004 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" (UID: "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.322742 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-config" (OuterVolumeSpecName: "config") pod "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" (UID: "9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.356135 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.356175 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.356191 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.356203 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.356212 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.377577 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.825973 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-584ffbf4c6-g7bkv"] Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.836164 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.840647 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.841011 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.841081 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.859773 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.860057 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.862486 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9h99t" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.920959 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-584ffbf4c6-g7bkv"] Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984293 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-scripts\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984388 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-config-data\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984470 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-combined-ca-bundle\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984517 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-internal-tls-certs\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984583 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-fernet-keys\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984620 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-credential-keys\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984652 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-public-tls-certs\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:12 crc kubenswrapper[4758]: I1203 17:19:12.984737 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-924wf\" (UniqueName: \"kubernetes.io/projected/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-kube-api-access-924wf\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.011068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ca64c8-28c7-472e-85bd-acd6af307580","Type":"ContainerStarted","Data":"ea6dcf9c2fc3a2c5e5cd9cbbf22a997692e23bc0e78284fb1d6304f44a979e57"} Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.032421 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"12d6c948-8253-4dbf-b782-27603b7ae061","Type":"ContainerStarted","Data":"cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d"} Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087513 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-fernet-keys\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087590 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-credential-keys\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087636 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-public-tls-certs\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087722 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-924wf\" (UniqueName: \"kubernetes.io/projected/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-kube-api-access-924wf\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087822 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-scripts\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087871 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-config-data\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.087986 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-combined-ca-bundle\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.088084 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-internal-tls-certs\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.096117 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-internal-tls-certs\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.111310 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-scripts\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.111919 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a"} Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.133936 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-fernet-keys\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.138001 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.137969796 podStartE2EDuration="11.137969796s" podCreationTimestamp="2025-12-03 17:19:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:13.085622562 +0000 UTC m=+1408.286999433" watchObservedRunningTime="2025-12-03 17:19:13.137969796 +0000 UTC m=+1408.339346657" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.148578 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-combined-ca-bundle\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.163823 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-public-tls-certs\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.178501 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-924wf\" (UniqueName: \"kubernetes.io/projected/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-kube-api-access-924wf\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.195986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-config-data\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.197936 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-credential-keys\") pod \"keystone-584ffbf4c6-g7bkv\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.210489 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4ggvx" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.214398 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerStarted","Data":"2bad2b11cca0d4a475d1f157b37f370905f6d7ad77235abfbb17cdd14e36cff6"} Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.237199 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.237268 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.276714 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.287969 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.315740 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68d78d4d4d-rhzqc"] Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.415608 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.435754 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4ggvx"] Dec 03 17:19:13 crc kubenswrapper[4758]: I1203 17:19:13.454189 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4ggvx"] Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.003269 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-584ffbf4c6-g7bkv"] Dec 03 17:19:14 crc kubenswrapper[4758]: W1203 17:19:14.019772 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bdefd9e_77f0_4137_8b30_35aca9f17c3f.slice/crio-be4a4e6fb13c4e45b0dc305c4994a606ae4bcc366eabeac00971b054e1a8939c WatchSource:0}: Error finding container be4a4e6fb13c4e45b0dc305c4994a606ae4bcc366eabeac00971b054e1a8939c: Status 404 returned error can't find the container with id be4a4e6fb13c4e45b0dc305c4994a606ae4bcc366eabeac00971b054e1a8939c Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.226120 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d78d4d4d-rhzqc" event={"ID":"1f328daf-ac93-432b-8645-14b7030a24d1","Type":"ContainerStarted","Data":"e9526364cd771e1d73395320b28267b014ffd81941f74ef5e036bbd181871def"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.226830 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.226852 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d78d4d4d-rhzqc" event={"ID":"1f328daf-ac93-432b-8645-14b7030a24d1","Type":"ContainerStarted","Data":"7ff34332c6d8c7344235f023f03dbde628713fc91c197634ffc70e74d0fc3ae7"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.226867 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d78d4d4d-rhzqc" event={"ID":"1f328daf-ac93-432b-8645-14b7030a24d1","Type":"ContainerStarted","Data":"d188692e4f7d254f836db7bf4c350e95d6c6d3c30b3458283cab4b30afa448da"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.228169 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"12d6c948-8253-4dbf-b782-27603b7ae061","Type":"ContainerStarted","Data":"93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.231285 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6l8jp" event={"ID":"dbfaa3a1-7e61-4537-b953-07269f386adc","Type":"ContainerStarted","Data":"ae9f6c0cce128861e191ebc593d6fa56a7d7fd6eeadf7eaf6a4c003584c87ce7"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.244661 4758 generic.go:334] "Generic (PLEG): container finished" podID="6b98f774-a8d0-460e-a718-c119dbd1f203" containerID="c4f33dad18cd24c6fb9262563f5df5def2e2efa7c561f64389ce9ac18f00ff1c" exitCode=0 Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.244895 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-q2vlf" event={"ID":"6b98f774-a8d0-460e-a718-c119dbd1f203","Type":"ContainerDied","Data":"c4f33dad18cd24c6fb9262563f5df5def2e2efa7c561f64389ce9ac18f00ff1c"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.249267 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-584ffbf4c6-g7bkv" event={"ID":"6bdefd9e-77f0-4137-8b30-35aca9f17c3f","Type":"ContainerStarted","Data":"be4a4e6fb13c4e45b0dc305c4994a606ae4bcc366eabeac00971b054e1a8939c"} Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.249848 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.249988 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.257127 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68d78d4d4d-rhzqc" podStartSLOduration=3.257094436 podStartE2EDuration="3.257094436s" podCreationTimestamp="2025-12-03 17:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:14.253792906 +0000 UTC m=+1409.455169767" watchObservedRunningTime="2025-12-03 17:19:14.257094436 +0000 UTC m=+1409.458471297" Dec 03 17:19:14 crc kubenswrapper[4758]: I1203 17:19:14.285105 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.285071891 podStartE2EDuration="11.285071891s" podCreationTimestamp="2025-12-03 17:19:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:14.283623482 +0000 UTC m=+1409.485000343" watchObservedRunningTime="2025-12-03 17:19:14.285071891 +0000 UTC m=+1409.486448752" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.132468 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9" path="/var/lib/kubelet/pods/9f4ec3d3-b3e2-426d-84bb-a4453d8d58a9/volumes" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.164487 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-6l8jp" podStartSLOduration=5.499992067 podStartE2EDuration="49.164418731s" podCreationTimestamp="2025-12-03 17:18:26 +0000 UTC" firstStartedPulling="2025-12-03 17:18:28.086708117 +0000 UTC m=+1363.288084978" lastFinishedPulling="2025-12-03 17:19:11.751134781 +0000 UTC m=+1406.952511642" observedRunningTime="2025-12-03 17:19:14.336911954 +0000 UTC m=+1409.538288815" watchObservedRunningTime="2025-12-03 17:19:15.164418731 +0000 UTC m=+1410.365795592" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.287037 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-584ffbf4c6-g7bkv" event={"ID":"6bdefd9e-77f0-4137-8b30-35aca9f17c3f","Type":"ContainerStarted","Data":"1d73cb8d75674ff7493a3d6896b8923851258760396fa1c116f8e006497ff57c"} Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.290042 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.290114 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.313934 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-584ffbf4c6-g7bkv" podStartSLOduration=3.31390211 podStartE2EDuration="3.31390211s" podCreationTimestamp="2025-12-03 17:19:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:15.313210512 +0000 UTC m=+1410.514587383" watchObservedRunningTime="2025-12-03 17:19:15.31390211 +0000 UTC m=+1410.515278971" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.801235 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.874305 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-combined-ca-bundle\") pod \"6b98f774-a8d0-460e-a718-c119dbd1f203\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.875096 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wzts\" (UniqueName: \"kubernetes.io/projected/6b98f774-a8d0-460e-a718-c119dbd1f203-kube-api-access-7wzts\") pod \"6b98f774-a8d0-460e-a718-c119dbd1f203\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.875406 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-db-sync-config-data\") pod \"6b98f774-a8d0-460e-a718-c119dbd1f203\" (UID: \"6b98f774-a8d0-460e-a718-c119dbd1f203\") " Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.882108 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b98f774-a8d0-460e-a718-c119dbd1f203-kube-api-access-7wzts" (OuterVolumeSpecName: "kube-api-access-7wzts") pod "6b98f774-a8d0-460e-a718-c119dbd1f203" (UID: "6b98f774-a8d0-460e-a718-c119dbd1f203"). InnerVolumeSpecName "kube-api-access-7wzts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.911263 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b98f774-a8d0-460e-a718-c119dbd1f203" (UID: "6b98f774-a8d0-460e-a718-c119dbd1f203"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.970446 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6b98f774-a8d0-460e-a718-c119dbd1f203" (UID: "6b98f774-a8d0-460e-a718-c119dbd1f203"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.978449 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wzts\" (UniqueName: \"kubernetes.io/projected/6b98f774-a8d0-460e-a718-c119dbd1f203-kube-api-access-7wzts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.978494 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:15 crc kubenswrapper[4758]: I1203 17:19:15.978504 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b98f774-a8d0-460e-a718-c119dbd1f203-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.299496 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-q2vlf" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.311333 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-q2vlf" event={"ID":"6b98f774-a8d0-460e-a718-c119dbd1f203","Type":"ContainerDied","Data":"6c16a1550f4aaad99325521243704066ddeea7e997248714a31a521d276af136"} Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.311396 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c16a1550f4aaad99325521243704066ddeea7e997248714a31a521d276af136" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.639355 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-57b5ff9967-k6s52"] Dec 03 17:19:16 crc kubenswrapper[4758]: E1203 17:19:16.640064 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b98f774-a8d0-460e-a718-c119dbd1f203" containerName="barbican-db-sync" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.640109 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b98f774-a8d0-460e-a718-c119dbd1f203" containerName="barbican-db-sync" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.640368 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b98f774-a8d0-460e-a718-c119dbd1f203" containerName="barbican-db-sync" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.642042 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.648173 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.648555 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-585nf" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.648607 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.691536 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7bb9966db6-phk6t"] Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.694049 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.696705 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-logs\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.696880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.696909 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data-custom\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.696967 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlsxt\" (UniqueName: \"kubernetes.io/projected/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-kube-api-access-dlsxt\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.697018 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-combined-ca-bundle\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.698290 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.706772 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57b5ff9967-k6s52"] Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.719208 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bb9966db6-phk6t"] Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.741851 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-45cpd"] Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.744843 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.775360 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-45cpd"] Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.804863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.804961 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data-custom\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805071 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-config\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805112 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data-custom\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805239 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlsxt\" (UniqueName: \"kubernetes.io/projected/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-kube-api-access-dlsxt\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805284 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805395 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-combined-ca-bundle\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805437 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805479 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkn7l\" (UniqueName: \"kubernetes.io/projected/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-kube-api-access-rkn7l\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.805546 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfhqs\" (UniqueName: \"kubernetes.io/projected/53b74dc9-5836-4c5c-bae8-803f64c9de12-kube-api-access-sfhqs\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.806832 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-logs\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.806928 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.807128 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.807216 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-combined-ca-bundle\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.807305 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.807491 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b74dc9-5836-4c5c-bae8-803f64c9de12-logs\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.815428 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-logs\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.831762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data-custom\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.837606 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-combined-ca-bundle\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.898565 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlsxt\" (UniqueName: \"kubernetes.io/projected/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-kube-api-access-dlsxt\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.905851 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data\") pod \"barbican-worker-57b5ff9967-k6s52\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.910622 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b74dc9-5836-4c5c-bae8-803f64c9de12-logs\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.910827 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-config\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.910884 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data-custom\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.910950 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911006 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911064 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkn7l\" (UniqueName: \"kubernetes.io/projected/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-kube-api-access-rkn7l\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911123 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfhqs\" (UniqueName: \"kubernetes.io/projected/53b74dc9-5836-4c5c-bae8-803f64c9de12-kube-api-access-sfhqs\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911245 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-combined-ca-bundle\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.911536 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.916066 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.916250 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-config\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.918661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.918837 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.919652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b74dc9-5836-4c5c-bae8-803f64c9de12-logs\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.920115 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.927360 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data-custom\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.928976 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.944622 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65c99cf6db-25flz"] Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.947224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.950246 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfhqs\" (UniqueName: \"kubernetes.io/projected/53b74dc9-5836-4c5c-bae8-803f64c9de12-kube-api-access-sfhqs\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.952971 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-combined-ca-bundle\") pod \"barbican-keystone-listener-7bb9966db6-phk6t\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.954235 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.960209 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkn7l\" (UniqueName: \"kubernetes.io/projected/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-kube-api-access-rkn7l\") pod \"dnsmasq-dns-7c67bffd47-45cpd\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:16 crc kubenswrapper[4758]: I1203 17:19:16.961418 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65c99cf6db-25flz"] Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.013274 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.014542 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-combined-ca-bundle\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.014603 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85bc93b-b75d-4d54-aa50-cc9027c0f044-logs\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.014762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t82rf\" (UniqueName: \"kubernetes.io/projected/b85bc93b-b75d-4d54-aa50-cc9027c0f044-kube-api-access-t82rf\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.015428 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data-custom\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.015453 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.044065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.122367 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.123494 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-combined-ca-bundle\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.123549 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85bc93b-b75d-4d54-aa50-cc9027c0f044-logs\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.124510 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85bc93b-b75d-4d54-aa50-cc9027c0f044-logs\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.125901 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t82rf\" (UniqueName: \"kubernetes.io/projected/b85bc93b-b75d-4d54-aa50-cc9027c0f044-kube-api-access-t82rf\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.125965 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.125985 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data-custom\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.132261 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data-custom\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.136103 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-combined-ca-bundle\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.145633 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t82rf\" (UniqueName: \"kubernetes.io/projected/b85bc93b-b75d-4d54-aa50-cc9027c0f044-kube-api-access-t82rf\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.149835 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data\") pod \"barbican-api-65c99cf6db-25flz\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.151312 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.336836 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.698198 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57b5ff9967-k6s52"] Dec 03 17:19:17 crc kubenswrapper[4758]: W1203 17:19:17.705414 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod218264a1_c6a9_4ecb_8f40_5df1fd7dc77c.slice/crio-48c44221348cf58dc317ed0742cadeb940a7fb2be358ee47a42d1c2dd8be5e84 WatchSource:0}: Error finding container 48c44221348cf58dc317ed0742cadeb940a7fb2be358ee47a42d1c2dd8be5e84: Status 404 returned error can't find the container with id 48c44221348cf58dc317ed0742cadeb940a7fb2be358ee47a42d1c2dd8be5e84 Dec 03 17:19:17 crc kubenswrapper[4758]: W1203 17:19:17.708538 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeeca44d2_0d4c_4b1e_957b_e51cf0fed653.slice/crio-5ed74a9e6011fb877963860f09a9ff62a7821fc5c7954bac8419e9e88aed78c2 WatchSource:0}: Error finding container 5ed74a9e6011fb877963860f09a9ff62a7821fc5c7954bac8419e9e88aed78c2: Status 404 returned error can't find the container with id 5ed74a9e6011fb877963860f09a9ff62a7821fc5c7954bac8419e9e88aed78c2 Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.713781 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-45cpd"] Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.912558 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bb9966db6-phk6t"] Dec 03 17:19:17 crc kubenswrapper[4758]: W1203 17:19:17.912604 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb85bc93b_b75d_4d54_aa50_cc9027c0f044.slice/crio-b5f64dc80b3325a499d2176db7720bba2ebd55e5cb0f3d970ab5965aed52954b WatchSource:0}: Error finding container b5f64dc80b3325a499d2176db7720bba2ebd55e5cb0f3d970ab5965aed52954b: Status 404 returned error can't find the container with id b5f64dc80b3325a499d2176db7720bba2ebd55e5cb0f3d970ab5965aed52954b Dec 03 17:19:17 crc kubenswrapper[4758]: I1203 17:19:17.926216 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65c99cf6db-25flz"] Dec 03 17:19:18 crc kubenswrapper[4758]: I1203 17:19:18.358468 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" event={"ID":"53b74dc9-5836-4c5c-bae8-803f64c9de12","Type":"ContainerStarted","Data":"89ab0ca141fa024b4e4a491501e9e620f3075821ae28fae8bcfdd12d52ef5bbd"} Dec 03 17:19:18 crc kubenswrapper[4758]: I1203 17:19:18.362392 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" event={"ID":"eeca44d2-0d4c-4b1e-957b-e51cf0fed653","Type":"ContainerStarted","Data":"5ed74a9e6011fb877963860f09a9ff62a7821fc5c7954bac8419e9e88aed78c2"} Dec 03 17:19:18 crc kubenswrapper[4758]: I1203 17:19:18.364519 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65c99cf6db-25flz" event={"ID":"b85bc93b-b75d-4d54-aa50-cc9027c0f044","Type":"ContainerStarted","Data":"b5f64dc80b3325a499d2176db7720bba2ebd55e5cb0f3d970ab5965aed52954b"} Dec 03 17:19:18 crc kubenswrapper[4758]: I1203 17:19:18.366365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57b5ff9967-k6s52" event={"ID":"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c","Type":"ContainerStarted","Data":"48c44221348cf58dc317ed0742cadeb940a7fb2be358ee47a42d1c2dd8be5e84"} Dec 03 17:19:19 crc kubenswrapper[4758]: I1203 17:19:19.412248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65c99cf6db-25flz" event={"ID":"b85bc93b-b75d-4d54-aa50-cc9027c0f044","Type":"ContainerStarted","Data":"edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e"} Dec 03 17:19:19 crc kubenswrapper[4758]: I1203 17:19:19.413108 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65c99cf6db-25flz" event={"ID":"b85bc93b-b75d-4d54-aa50-cc9027c0f044","Type":"ContainerStarted","Data":"4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f"} Dec 03 17:19:19 crc kubenswrapper[4758]: I1203 17:19:19.429621 4758 generic.go:334] "Generic (PLEG): container finished" podID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerID="c52488b1fd8bbe17a364f336ed8c9e4a535697f826f228181c3b08a671b1af91" exitCode=0 Dec 03 17:19:19 crc kubenswrapper[4758]: I1203 17:19:19.429721 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" event={"ID":"eeca44d2-0d4c-4b1e-957b-e51cf0fed653","Type":"ContainerDied","Data":"c52488b1fd8bbe17a364f336ed8c9e4a535697f826f228181c3b08a671b1af91"} Dec 03 17:19:19 crc kubenswrapper[4758]: I1203 17:19:19.500891 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.111134 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-576b5fdbd4-qmczw"] Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.114065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.117848 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.118139 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.136073 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-576b5fdbd4-qmczw"] Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.257931 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-combined-ca-bundle\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.258010 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-internal-tls-certs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.258128 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-logs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.258171 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.258198 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-public-tls-certs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.258704 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87nfc\" (UniqueName: \"kubernetes.io/projected/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-kube-api-access-87nfc\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.259937 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data-custom\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.362518 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-combined-ca-bundle\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.362576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-internal-tls-certs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.362648 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-logs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.362695 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.363214 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-logs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.363299 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-public-tls-certs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.363870 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87nfc\" (UniqueName: \"kubernetes.io/projected/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-kube-api-access-87nfc\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.364072 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data-custom\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.372273 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data-custom\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.374082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-combined-ca-bundle\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.375378 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-public-tls-certs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.379630 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.382841 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-internal-tls-certs\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.385861 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87nfc\" (UniqueName: \"kubernetes.io/projected/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-kube-api-access-87nfc\") pod \"barbican-api-576b5fdbd4-qmczw\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.448535 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.448634 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.453418 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:20 crc kubenswrapper[4758]: I1203 17:19:20.484698 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65c99cf6db-25flz" podStartSLOduration=4.484651193 podStartE2EDuration="4.484651193s" podCreationTimestamp="2025-12-03 17:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:20.483173303 +0000 UTC m=+1415.684550174" watchObservedRunningTime="2025-12-03 17:19:20.484651193 +0000 UTC m=+1415.686028054" Dec 03 17:19:24 crc kubenswrapper[4758]: I1203 17:19:24.184024 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:24 crc kubenswrapper[4758]: I1203 17:19:24.184889 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:24 crc kubenswrapper[4758]: I1203 17:19:24.231918 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:24 crc kubenswrapper[4758]: I1203 17:19:24.235822 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:24 crc kubenswrapper[4758]: I1203 17:19:24.503237 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:24 crc kubenswrapper[4758]: I1203 17:19:24.503286 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:26 crc kubenswrapper[4758]: I1203 17:19:26.530880 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:19:26 crc kubenswrapper[4758]: I1203 17:19:26.531345 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:19:26 crc kubenswrapper[4758]: I1203 17:19:26.801807 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:26 crc kubenswrapper[4758]: I1203 17:19:26.864559 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 17:19:27 crc kubenswrapper[4758]: I1203 17:19:27.750083 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-576b5fdbd4-qmczw"] Dec 03 17:19:27 crc kubenswrapper[4758]: E1203 17:19:27.982952 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.567158 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerStarted","Data":"19b56edda648ad38df5de5fb1958faf6c38b399f4a2cdaea93793667dcf943e2"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.567423 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="ceilometer-notification-agent" containerID="cri-o://e28be1b4932b60511077a9023fc129e0953206f967f40a629526334ce7f94d00" gracePeriod=30 Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.567888 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.568293 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="proxy-httpd" containerID="cri-o://19b56edda648ad38df5de5fb1958faf6c38b399f4a2cdaea93793667dcf943e2" gracePeriod=30 Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.568387 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="sg-core" containerID="cri-o://2bad2b11cca0d4a475d1f157b37f370905f6d7ad77235abfbb17cdd14e36cff6" gracePeriod=30 Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.571223 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57b5ff9967-k6s52" event={"ID":"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c","Type":"ContainerStarted","Data":"f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.571266 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57b5ff9967-k6s52" event={"ID":"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c","Type":"ContainerStarted","Data":"7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.580287 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-576b5fdbd4-qmczw" event={"ID":"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665","Type":"ContainerStarted","Data":"c1e072acd4c62b59d1cb154ec06ad176f5ed4d1bdd3cc652119c966d7118b7f9"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.580363 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-576b5fdbd4-qmczw" event={"ID":"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665","Type":"ContainerStarted","Data":"99106e146935a884c8eae3a30d8b26d2549f367f4b50ca0f4536926fc866b59a"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.580377 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-576b5fdbd4-qmczw" event={"ID":"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665","Type":"ContainerStarted","Data":"1a30804cb4612cabd8244c764409e4ecfb725677392161471dcd1712c9a39290"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.580785 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.580831 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.585160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" event={"ID":"eeca44d2-0d4c-4b1e-957b-e51cf0fed653","Type":"ContainerStarted","Data":"39a064a7ed198a4f95027678b146f34a652fa2a1c6669661643c18587b1970f0"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.590655 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" event={"ID":"53b74dc9-5836-4c5c-bae8-803f64c9de12","Type":"ContainerStarted","Data":"efb795a450acef7b396db339028440c950cbfa2f32b9d7f9c051e83e71c85fb6"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.590738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" event={"ID":"53b74dc9-5836-4c5c-bae8-803f64c9de12","Type":"ContainerStarted","Data":"9c02a3775a91ddea9562f2c00063105c43fac3fa4443572b50d1f8eb0a7c2803"} Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.639714 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-57b5ff9967-k6s52" podStartSLOduration=3.17741271 podStartE2EDuration="12.639666699s" podCreationTimestamp="2025-12-03 17:19:16 +0000 UTC" firstStartedPulling="2025-12-03 17:19:17.70866168 +0000 UTC m=+1412.910038541" lastFinishedPulling="2025-12-03 17:19:27.170915669 +0000 UTC m=+1422.372292530" observedRunningTime="2025-12-03 17:19:28.626125393 +0000 UTC m=+1423.827502254" watchObservedRunningTime="2025-12-03 17:19:28.639666699 +0000 UTC m=+1423.841043560" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.664940 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" podStartSLOduration=3.4145552 podStartE2EDuration="12.664900591s" podCreationTimestamp="2025-12-03 17:19:16 +0000 UTC" firstStartedPulling="2025-12-03 17:19:17.91880558 +0000 UTC m=+1413.120182441" lastFinishedPulling="2025-12-03 17:19:27.169150971 +0000 UTC m=+1422.370527832" observedRunningTime="2025-12-03 17:19:28.650103501 +0000 UTC m=+1423.851480372" watchObservedRunningTime="2025-12-03 17:19:28.664900591 +0000 UTC m=+1423.866277452" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.699751 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-576b5fdbd4-qmczw" podStartSLOduration=8.699725962 podStartE2EDuration="8.699725962s" podCreationTimestamp="2025-12-03 17:19:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:28.698309494 +0000 UTC m=+1423.899686355" watchObservedRunningTime="2025-12-03 17:19:28.699725962 +0000 UTC m=+1423.901102823" Dec 03 17:19:28 crc kubenswrapper[4758]: I1203 17:19:28.751902 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" podStartSLOduration=12.751878332 podStartE2EDuration="12.751878332s" podCreationTimestamp="2025-12-03 17:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:28.741386068 +0000 UTC m=+1423.942762949" watchObservedRunningTime="2025-12-03 17:19:28.751878332 +0000 UTC m=+1423.953255193" Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.308457 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.603546 4758 generic.go:334] "Generic (PLEG): container finished" podID="dbfaa3a1-7e61-4537-b953-07269f386adc" containerID="ae9f6c0cce128861e191ebc593d6fa56a7d7fd6eeadf7eaf6a4c003584c87ce7" exitCode=0 Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.603658 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6l8jp" event={"ID":"dbfaa3a1-7e61-4537-b953-07269f386adc","Type":"ContainerDied","Data":"ae9f6c0cce128861e191ebc593d6fa56a7d7fd6eeadf7eaf6a4c003584c87ce7"} Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.607950 4758 generic.go:334] "Generic (PLEG): container finished" podID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerID="19b56edda648ad38df5de5fb1958faf6c38b399f4a2cdaea93793667dcf943e2" exitCode=0 Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.608002 4758 generic.go:334] "Generic (PLEG): container finished" podID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerID="2bad2b11cca0d4a475d1f157b37f370905f6d7ad77235abfbb17cdd14e36cff6" exitCode=2 Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.608000 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerDied","Data":"19b56edda648ad38df5de5fb1958faf6c38b399f4a2cdaea93793667dcf943e2"} Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.608079 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerDied","Data":"2bad2b11cca0d4a475d1f157b37f370905f6d7ad77235abfbb17cdd14e36cff6"} Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.609539 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:29 crc kubenswrapper[4758]: I1203 17:19:29.905506 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.626266 4758 generic.go:334] "Generic (PLEG): container finished" podID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerID="e28be1b4932b60511077a9023fc129e0953206f967f40a629526334ce7f94d00" exitCode=0 Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.626350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerDied","Data":"e28be1b4932b60511077a9023fc129e0953206f967f40a629526334ce7f94d00"} Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.629535 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"994006af-c89c-42b1-b0b7-f6a247e83cab","Type":"ContainerDied","Data":"5933d3899283877952151052a4c96938c455a4b0a66e63b9e69fea0440546262"} Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.629606 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5933d3899283877952151052a4c96938c455a4b0a66e63b9e69fea0440546262" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.765341 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891277 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-scripts\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891362 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-config-data\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891584 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-run-httpd\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891639 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj57t\" (UniqueName: \"kubernetes.io/projected/994006af-c89c-42b1-b0b7-f6a247e83cab-kube-api-access-sj57t\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891715 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-sg-core-conf-yaml\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891806 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-combined-ca-bundle\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.891854 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-log-httpd\") pod \"994006af-c89c-42b1-b0b7-f6a247e83cab\" (UID: \"994006af-c89c-42b1-b0b7-f6a247e83cab\") " Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.893538 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.893826 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.916878 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/994006af-c89c-42b1-b0b7-f6a247e83cab-kube-api-access-sj57t" (OuterVolumeSpecName: "kube-api-access-sj57t") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "kube-api-access-sj57t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.916932 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-scripts" (OuterVolumeSpecName: "scripts") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.959281 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.976040 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.995713 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.995764 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.995775 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.995787 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/994006af-c89c-42b1-b0b7-f6a247e83cab-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.995796 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj57t\" (UniqueName: \"kubernetes.io/projected/994006af-c89c-42b1-b0b7-f6a247e83cab-kube-api-access-sj57t\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.995815 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:30 crc kubenswrapper[4758]: I1203 17:19:30.998339 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-config-data" (OuterVolumeSpecName: "config-data") pod "994006af-c89c-42b1-b0b7-f6a247e83cab" (UID: "994006af-c89c-42b1-b0b7-f6a247e83cab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.038186 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.097698 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbfaa3a1-7e61-4537-b953-07269f386adc-etc-machine-id\") pod \"dbfaa3a1-7e61-4537-b953-07269f386adc\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.097778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-db-sync-config-data\") pod \"dbfaa3a1-7e61-4537-b953-07269f386adc\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.097931 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cl5m\" (UniqueName: \"kubernetes.io/projected/dbfaa3a1-7e61-4537-b953-07269f386adc-kube-api-access-2cl5m\") pod \"dbfaa3a1-7e61-4537-b953-07269f386adc\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.097987 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-scripts\") pod \"dbfaa3a1-7e61-4537-b953-07269f386adc\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.098023 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-combined-ca-bundle\") pod \"dbfaa3a1-7e61-4537-b953-07269f386adc\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.098074 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-config-data\") pod \"dbfaa3a1-7e61-4537-b953-07269f386adc\" (UID: \"dbfaa3a1-7e61-4537-b953-07269f386adc\") " Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.098624 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994006af-c89c-42b1-b0b7-f6a247e83cab-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.099169 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dbfaa3a1-7e61-4537-b953-07269f386adc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dbfaa3a1-7e61-4537-b953-07269f386adc" (UID: "dbfaa3a1-7e61-4537-b953-07269f386adc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.103875 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-scripts" (OuterVolumeSpecName: "scripts") pod "dbfaa3a1-7e61-4537-b953-07269f386adc" (UID: "dbfaa3a1-7e61-4537-b953-07269f386adc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.104478 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dbfaa3a1-7e61-4537-b953-07269f386adc" (UID: "dbfaa3a1-7e61-4537-b953-07269f386adc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.116870 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbfaa3a1-7e61-4537-b953-07269f386adc-kube-api-access-2cl5m" (OuterVolumeSpecName: "kube-api-access-2cl5m") pod "dbfaa3a1-7e61-4537-b953-07269f386adc" (UID: "dbfaa3a1-7e61-4537-b953-07269f386adc"). InnerVolumeSpecName "kube-api-access-2cl5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.136004 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbfaa3a1-7e61-4537-b953-07269f386adc" (UID: "dbfaa3a1-7e61-4537-b953-07269f386adc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.155883 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-config-data" (OuterVolumeSpecName: "config-data") pod "dbfaa3a1-7e61-4537-b953-07269f386adc" (UID: "dbfaa3a1-7e61-4537-b953-07269f386adc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.202575 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.202897 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.202996 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.203062 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbfaa3a1-7e61-4537-b953-07269f386adc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.203124 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbfaa3a1-7e61-4537-b953-07269f386adc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.203187 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cl5m\" (UniqueName: \"kubernetes.io/projected/dbfaa3a1-7e61-4537-b953-07269f386adc-kube-api-access-2cl5m\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.638417 4758 generic.go:334] "Generic (PLEG): container finished" podID="e6bb6321-e853-40ce-a5f6-7b7a53e26595" containerID="7247d58e71f563549d53c42917cec0911f2f98ae88698a1caa308986151490b6" exitCode=0 Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.638492 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j7q7w" event={"ID":"e6bb6321-e853-40ce-a5f6-7b7a53e26595","Type":"ContainerDied","Data":"7247d58e71f563549d53c42917cec0911f2f98ae88698a1caa308986151490b6"} Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.641279 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.642467 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6l8jp" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.646307 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6l8jp" event={"ID":"dbfaa3a1-7e61-4537-b953-07269f386adc","Type":"ContainerDied","Data":"fce8c67cc946cbd7a2cbffe68752bb07c36640b8f446577a94c9f496ea5c28df"} Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.646598 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fce8c67cc946cbd7a2cbffe68752bb07c36640b8f446577a94c9f496ea5c28df" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.717400 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.735759 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.750933 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:19:31 crc kubenswrapper[4758]: E1203 17:19:31.751615 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="ceilometer-notification-agent" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.751637 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="ceilometer-notification-agent" Dec 03 17:19:31 crc kubenswrapper[4758]: E1203 17:19:31.751669 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbfaa3a1-7e61-4537-b953-07269f386adc" containerName="cinder-db-sync" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.751697 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbfaa3a1-7e61-4537-b953-07269f386adc" containerName="cinder-db-sync" Dec 03 17:19:31 crc kubenswrapper[4758]: E1203 17:19:31.751732 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="proxy-httpd" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.751742 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="proxy-httpd" Dec 03 17:19:31 crc kubenswrapper[4758]: E1203 17:19:31.751761 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="sg-core" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.751769 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="sg-core" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.752034 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="sg-core" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.752055 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbfaa3a1-7e61-4537-b953-07269f386adc" containerName="cinder-db-sync" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.752086 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="ceilometer-notification-agent" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.752101 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" containerName="proxy-httpd" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.755970 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.759437 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.759736 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.778343 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.817895 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gphk4\" (UniqueName: \"kubernetes.io/projected/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-kube-api-access-gphk4\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.818113 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-log-httpd\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.818224 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.818277 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-config-data\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.818358 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-run-httpd\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.818617 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-scripts\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.818725 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.918586 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.920926 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-run-httpd\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.920995 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921058 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-scripts\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921110 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921161 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gphk4\" (UniqueName: \"kubernetes.io/projected/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-kube-api-access-gphk4\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921199 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-log-httpd\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921262 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-config-data\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.921536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-run-httpd\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.922489 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-log-httpd\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.929883 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.930340 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-scripts\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.931715 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.931773 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.931946 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.931715 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5fmvm" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.939134 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-config-data\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.939511 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.953578 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gphk4\" (UniqueName: \"kubernetes.io/projected/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-kube-api-access-gphk4\") pod \"ceilometer-0\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " pod="openstack/ceilometer-0" Dec 03 17:19:31 crc kubenswrapper[4758]: I1203 17:19:31.964381 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.027234 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.027300 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.027384 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kmd8\" (UniqueName: \"kubernetes.io/projected/042060a5-b97e-49a0-9d5f-f946c785cdf3-kube-api-access-9kmd8\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.027440 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.027552 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/042060a5-b97e-49a0-9d5f-f946c785cdf3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.027622 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-scripts\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.079301 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.129805 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-45cpd"] Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.130908 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-scripts\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.138261 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.138399 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.138506 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kmd8\" (UniqueName: \"kubernetes.io/projected/042060a5-b97e-49a0-9d5f-f946c785cdf3-kube-api-access-9kmd8\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.138590 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.138850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/042060a5-b97e-49a0-9d5f-f946c785cdf3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.139101 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/042060a5-b97e-49a0-9d5f-f946c785cdf3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.139661 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerName="dnsmasq-dns" containerID="cri-o://39a064a7ed198a4f95027678b146f34a652fa2a1c6669661643c18587b1970f0" gracePeriod=10 Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.147851 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.155276 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.169875 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-scripts\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.170114 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.180919 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.185573 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kmd8\" (UniqueName: \"kubernetes.io/projected/042060a5-b97e-49a0-9d5f-f946c785cdf3-kube-api-access-9kmd8\") pod \"cinder-scheduler-0\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.211594 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj"] Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.223755 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.230313 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj"] Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.245455 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-config\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.245784 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.245898 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs6pl\" (UniqueName: \"kubernetes.io/projected/2878a24c-651b-46dc-87fd-78243a14aae0-kube-api-access-vs6pl\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.245942 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.245985 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.246016 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-svc\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.349963 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.352701 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.356508 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.359783 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs6pl\" (UniqueName: \"kubernetes.io/projected/2878a24c-651b-46dc-87fd-78243a14aae0-kube-api-access-vs6pl\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.359882 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.359961 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.360028 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-svc\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.360228 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-config\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.361453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.361592 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-svc\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.363223 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.386986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-config\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.400855 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs6pl\" (UniqueName: \"kubernetes.io/projected/2878a24c-651b-46dc-87fd-78243a14aae0-kube-api-access-vs6pl\") pod \"dnsmasq-dns-5cc8b5d5c5-k9qhj\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.520959 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.523371 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.529443 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.558269 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.587125 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data-custom\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.587247 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-scripts\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.587329 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-logs\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.587615 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.587735 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.587874 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.588020 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgfwm\" (UniqueName: \"kubernetes.io/projected/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-kube-api-access-dgfwm\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.666802 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.689694 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-scripts\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.689798 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-logs\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.689867 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.689912 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.689965 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.690030 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgfwm\" (UniqueName: \"kubernetes.io/projected/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-kube-api-access-dgfwm\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.690104 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data-custom\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.692382 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.692564 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-logs\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.702869 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.704035 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-scripts\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.707018 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.709997 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data-custom\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.713562 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgfwm\" (UniqueName: \"kubernetes.io/projected/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-kube-api-access-dgfwm\") pod \"cinder-api-0\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.715102 4758 generic.go:334] "Generic (PLEG): container finished" podID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerID="39a064a7ed198a4f95027678b146f34a652fa2a1c6669661643c18587b1970f0" exitCode=0 Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.715182 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" event={"ID":"eeca44d2-0d4c-4b1e-957b-e51cf0fed653","Type":"ContainerDied","Data":"39a064a7ed198a4f95027678b146f34a652fa2a1c6669661643c18587b1970f0"} Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.828921 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.894498 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-svc\") pod \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.894596 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-swift-storage-0\") pod \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.894650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-config\") pod \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.894790 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-sb\") pod \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.894842 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-nb\") pod \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.894907 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkn7l\" (UniqueName: \"kubernetes.io/projected/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-kube-api-access-rkn7l\") pod \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\" (UID: \"eeca44d2-0d4c-4b1e-957b-e51cf0fed653\") " Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.900911 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-kube-api-access-rkn7l" (OuterVolumeSpecName: "kube-api-access-rkn7l") pod "eeca44d2-0d4c-4b1e-957b-e51cf0fed653" (UID: "eeca44d2-0d4c-4b1e-957b-e51cf0fed653"). InnerVolumeSpecName "kube-api-access-rkn7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.924393 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.965525 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eeca44d2-0d4c-4b1e-957b-e51cf0fed653" (UID: "eeca44d2-0d4c-4b1e-957b-e51cf0fed653"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.965636 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eeca44d2-0d4c-4b1e-957b-e51cf0fed653" (UID: "eeca44d2-0d4c-4b1e-957b-e51cf0fed653"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.979191 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eeca44d2-0d4c-4b1e-957b-e51cf0fed653" (UID: "eeca44d2-0d4c-4b1e-957b-e51cf0fed653"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:32 crc kubenswrapper[4758]: I1203 17:19:32.985252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-config" (OuterVolumeSpecName: "config") pod "eeca44d2-0d4c-4b1e-957b-e51cf0fed653" (UID: "eeca44d2-0d4c-4b1e-957b-e51cf0fed653"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:32.998429 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:32.998472 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:32.998487 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:32.998497 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:32.998506 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkn7l\" (UniqueName: \"kubernetes.io/projected/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-kube-api-access-rkn7l\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.002609 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.022844 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.023948 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eeca44d2-0d4c-4b1e-957b-e51cf0fed653" (UID: "eeca44d2-0d4c-4b1e-957b-e51cf0fed653"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.106617 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeca44d2-0d4c-4b1e-957b-e51cf0fed653-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: W1203 17:19:33.121403 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod042060a5_b97e_49a0_9d5f_f946c785cdf3.slice/crio-d2942bfce12e33f4021f5395fed3366141da6a143cf6af0bdaf0b1ffa20dc224 WatchSource:0}: Error finding container d2942bfce12e33f4021f5395fed3366141da6a143cf6af0bdaf0b1ffa20dc224: Status 404 returned error can't find the container with id d2942bfce12e33f4021f5395fed3366141da6a143cf6af0bdaf0b1ffa20dc224 Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.133007 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="994006af-c89c-42b1-b0b7-f6a247e83cab" path="/var/lib/kubelet/pods/994006af-c89c-42b1-b0b7-f6a247e83cab/volumes" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.134263 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.150821 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.210488 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjcm\" (UniqueName: \"kubernetes.io/projected/e6bb6321-e853-40ce-a5f6-7b7a53e26595-kube-api-access-2tjcm\") pod \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.210698 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-combined-ca-bundle\") pod \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.210830 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-config\") pod \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\" (UID: \"e6bb6321-e853-40ce-a5f6-7b7a53e26595\") " Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.220580 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6bb6321-e853-40ce-a5f6-7b7a53e26595-kube-api-access-2tjcm" (OuterVolumeSpecName: "kube-api-access-2tjcm") pod "e6bb6321-e853-40ce-a5f6-7b7a53e26595" (UID: "e6bb6321-e853-40ce-a5f6-7b7a53e26595"). InnerVolumeSpecName "kube-api-access-2tjcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.240877 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-config" (OuterVolumeSpecName: "config") pod "e6bb6321-e853-40ce-a5f6-7b7a53e26595" (UID: "e6bb6321-e853-40ce-a5f6-7b7a53e26595"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.246407 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6bb6321-e853-40ce-a5f6-7b7a53e26595" (UID: "e6bb6321-e853-40ce-a5f6-7b7a53e26595"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.313658 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.313720 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e6bb6321-e853-40ce-a5f6-7b7a53e26595-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.313730 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjcm\" (UniqueName: \"kubernetes.io/projected/e6bb6321-e853-40ce-a5f6-7b7a53e26595-kube-api-access-2tjcm\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.351229 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj"] Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.486385 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:33 crc kubenswrapper[4758]: W1203 17:19:33.493157 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd65cf16_26a7_4dc2_b39e_23bc4afcb560.slice/crio-2b2d1235f2864db4211d0d34d838174306fb60db187db0a533cbd729a18d35e1 WatchSource:0}: Error finding container 2b2d1235f2864db4211d0d34d838174306fb60db187db0a533cbd729a18d35e1: Status 404 returned error can't find the container with id 2b2d1235f2864db4211d0d34d838174306fb60db187db0a533cbd729a18d35e1 Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.733075 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"042060a5-b97e-49a0-9d5f-f946c785cdf3","Type":"ContainerStarted","Data":"d2942bfce12e33f4021f5395fed3366141da6a143cf6af0bdaf0b1ffa20dc224"} Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.736254 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd65cf16-26a7-4dc2-b39e-23bc4afcb560","Type":"ContainerStarted","Data":"2b2d1235f2864db4211d0d34d838174306fb60db187db0a533cbd729a18d35e1"} Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.737808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerStarted","Data":"a393b197db8b6dfd4f7659361896c2477cdabc5dc2fe07f41dc02bddc2c7477a"} Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.739785 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" event={"ID":"2878a24c-651b-46dc-87fd-78243a14aae0","Type":"ContainerStarted","Data":"d1e6dd07e5bd6946b1435e2e5b1abb1ac7093addb7023b6557fe04bfbac48d2e"} Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.745812 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" event={"ID":"eeca44d2-0d4c-4b1e-957b-e51cf0fed653","Type":"ContainerDied","Data":"5ed74a9e6011fb877963860f09a9ff62a7821fc5c7954bac8419e9e88aed78c2"} Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.745803 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-45cpd" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.745916 4758 scope.go:117] "RemoveContainer" containerID="39a064a7ed198a4f95027678b146f34a652fa2a1c6669661643c18587b1970f0" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.750554 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j7q7w" event={"ID":"e6bb6321-e853-40ce-a5f6-7b7a53e26595","Type":"ContainerDied","Data":"36cc319da74db7336be5004041fcbc892adc19c3b44896ceb1971c901ccccbcb"} Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.750586 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36cc319da74db7336be5004041fcbc892adc19c3b44896ceb1971c901ccccbcb" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.750697 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j7q7w" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.789423 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-45cpd"] Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.816810 4758 scope.go:117] "RemoveContainer" containerID="c52488b1fd8bbe17a364f336ed8c9e4a535697f826f228181c3b08a671b1af91" Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.817066 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-45cpd"] Dec 03 17:19:33 crc kubenswrapper[4758]: I1203 17:19:33.991016 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj"] Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.074469 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7mn8j"] Dec 03 17:19:34 crc kubenswrapper[4758]: E1203 17:19:34.075002 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bb6321-e853-40ce-a5f6-7b7a53e26595" containerName="neutron-db-sync" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.075025 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bb6321-e853-40ce-a5f6-7b7a53e26595" containerName="neutron-db-sync" Dec 03 17:19:34 crc kubenswrapper[4758]: E1203 17:19:34.075037 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerName="dnsmasq-dns" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.075043 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerName="dnsmasq-dns" Dec 03 17:19:34 crc kubenswrapper[4758]: E1203 17:19:34.075089 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerName="init" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.075097 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerName="init" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.075281 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" containerName="dnsmasq-dns" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.075307 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bb6321-e853-40ce-a5f6-7b7a53e26595" containerName="neutron-db-sync" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.076655 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.137350 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rtxs\" (UniqueName: \"kubernetes.io/projected/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-kube-api-access-6rtxs\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.137418 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.137442 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.137473 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.137594 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-config\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.137617 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.146127 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7mn8j"] Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.199342 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64f4d5ff96-vfvv7"] Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.239745 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.240962 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-config\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.251302 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.251737 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rtxs\" (UniqueName: \"kubernetes.io/projected/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-kube-api-access-6rtxs\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.251870 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.252069 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.253913 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.247535 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.256383 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.262948 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.266048 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-config\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.274031 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.280065 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tt8c6" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.301123 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.301490 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.322431 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rtxs\" (UniqueName: \"kubernetes.io/projected/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-kube-api-access-6rtxs\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.344810 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7mn8j\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.378014 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-httpd-config\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.378412 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-combined-ca-bundle\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.378523 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvcsb\" (UniqueName: \"kubernetes.io/projected/d0887aea-e60c-4517-9c8c-c96498f912a4-kube-api-access-dvcsb\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.378572 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-config\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.378633 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-ovndb-tls-certs\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.404365 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64f4d5ff96-vfvv7"] Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.457051 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.487784 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-config\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.487858 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-ovndb-tls-certs\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.487933 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-httpd-config\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.488060 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-combined-ca-bundle\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.488105 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvcsb\" (UniqueName: \"kubernetes.io/projected/d0887aea-e60c-4517-9c8c-c96498f912a4-kube-api-access-dvcsb\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.499056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-config\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.501400 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-httpd-config\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.503269 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-ovndb-tls-certs\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.507494 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-combined-ca-bundle\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.514447 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvcsb\" (UniqueName: \"kubernetes.io/projected/d0887aea-e60c-4517-9c8c-c96498f912a4-kube-api-access-dvcsb\") pod \"neutron-64f4d5ff96-vfvv7\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.776586 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerStarted","Data":"d4fdb1467bc7e1d7119c3cd23b4a6f9c252938f343e72f940f4800aed1ad0cd9"} Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.783639 4758 generic.go:334] "Generic (PLEG): container finished" podID="2878a24c-651b-46dc-87fd-78243a14aae0" containerID="1fd2770e1f96ffa4829ca06694b0d06587aec363e8215c0195582962d2b0ff01" exitCode=0 Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.783755 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" event={"ID":"2878a24c-651b-46dc-87fd-78243a14aae0","Type":"ContainerDied","Data":"1fd2770e1f96ffa4829ca06694b0d06587aec363e8215c0195582962d2b0ff01"} Dec 03 17:19:34 crc kubenswrapper[4758]: I1203 17:19:34.787761 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.146672 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeca44d2-0d4c-4b1e-957b-e51cf0fed653" path="/var/lib/kubelet/pods/eeca44d2-0d4c-4b1e-957b-e51cf0fed653/volumes" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.448519 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.547257 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.562310 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7mn8j"] Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.753170 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-config\") pod \"2878a24c-651b-46dc-87fd-78243a14aae0\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.753283 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-swift-storage-0\") pod \"2878a24c-651b-46dc-87fd-78243a14aae0\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.753337 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-sb\") pod \"2878a24c-651b-46dc-87fd-78243a14aae0\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.753390 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs6pl\" (UniqueName: \"kubernetes.io/projected/2878a24c-651b-46dc-87fd-78243a14aae0-kube-api-access-vs6pl\") pod \"2878a24c-651b-46dc-87fd-78243a14aae0\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.753455 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-svc\") pod \"2878a24c-651b-46dc-87fd-78243a14aae0\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.753516 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-nb\") pod \"2878a24c-651b-46dc-87fd-78243a14aae0\" (UID: \"2878a24c-651b-46dc-87fd-78243a14aae0\") " Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.782542 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2878a24c-651b-46dc-87fd-78243a14aae0-kube-api-access-vs6pl" (OuterVolumeSpecName: "kube-api-access-vs6pl") pod "2878a24c-651b-46dc-87fd-78243a14aae0" (UID: "2878a24c-651b-46dc-87fd-78243a14aae0"). InnerVolumeSpecName "kube-api-access-vs6pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.790784 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2878a24c-651b-46dc-87fd-78243a14aae0" (UID: "2878a24c-651b-46dc-87fd-78243a14aae0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.791306 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2878a24c-651b-46dc-87fd-78243a14aae0" (UID: "2878a24c-651b-46dc-87fd-78243a14aae0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.829444 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2878a24c-651b-46dc-87fd-78243a14aae0" (UID: "2878a24c-651b-46dc-87fd-78243a14aae0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.845467 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2878a24c-651b-46dc-87fd-78243a14aae0" (UID: "2878a24c-651b-46dc-87fd-78243a14aae0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.861093 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd65cf16-26a7-4dc2-b39e-23bc4afcb560","Type":"ContainerStarted","Data":"a02e8862636694ec285582045f7961338abf66bf4ea94c2855817fa9605202b0"} Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.865289 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.865331 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.865345 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs6pl\" (UniqueName: \"kubernetes.io/projected/2878a24c-651b-46dc-87fd-78243a14aae0-kube-api-access-vs6pl\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.865356 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.865366 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.880628 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-config" (OuterVolumeSpecName: "config") pod "2878a24c-651b-46dc-87fd-78243a14aae0" (UID: "2878a24c-651b-46dc-87fd-78243a14aae0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.883477 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerStarted","Data":"9299b9e27e5198daa91a90f721e702d0ea26aa93e94e1904454efd0443ccf678"} Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.905203 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" event={"ID":"2878a24c-651b-46dc-87fd-78243a14aae0","Type":"ContainerDied","Data":"d1e6dd07e5bd6946b1435e2e5b1abb1ac7093addb7023b6557fe04bfbac48d2e"} Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.905294 4758 scope.go:117] "RemoveContainer" containerID="1fd2770e1f96ffa4829ca06694b0d06587aec363e8215c0195582962d2b0ff01" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.905537 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj" Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.924797 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" event={"ID":"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf","Type":"ContainerStarted","Data":"9d0f04cb2200c4228efbb032998e4ed185c43285269a35f43d01c9bb3cf5a985"} Dec 03 17:19:35 crc kubenswrapper[4758]: I1203 17:19:35.972290 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2878a24c-651b-46dc-87fd-78243a14aae0-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:36 crc kubenswrapper[4758]: I1203 17:19:36.071138 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj"] Dec 03 17:19:36 crc kubenswrapper[4758]: I1203 17:19:36.103795 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-k9qhj"] Dec 03 17:19:36 crc kubenswrapper[4758]: I1203 17:19:36.171735 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64f4d5ff96-vfvv7"] Dec 03 17:19:36 crc kubenswrapper[4758]: W1203 17:19:36.307396 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0887aea_e60c_4517_9c8c_c96498f912a4.slice/crio-8681e2c61ea1c18864ee8448a2f74d8ac7cc45bef5401d9e5ac9eaa8c1babd5d WatchSource:0}: Error finding container 8681e2c61ea1c18864ee8448a2f74d8ac7cc45bef5401d9e5ac9eaa8c1babd5d: Status 404 returned error can't find the container with id 8681e2c61ea1c18864ee8448a2f74d8ac7cc45bef5401d9e5ac9eaa8c1babd5d Dec 03 17:19:37 crc kubenswrapper[4758]: I1203 17:19:37.002557 4758 generic.go:334] "Generic (PLEG): container finished" podID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerID="96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16" exitCode=0 Dec 03 17:19:37 crc kubenswrapper[4758]: I1203 17:19:37.003098 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" event={"ID":"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf","Type":"ContainerDied","Data":"96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16"} Dec 03 17:19:37 crc kubenswrapper[4758]: I1203 17:19:37.009504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f4d5ff96-vfvv7" event={"ID":"d0887aea-e60c-4517-9c8c-c96498f912a4","Type":"ContainerStarted","Data":"8681e2c61ea1c18864ee8448a2f74d8ac7cc45bef5401d9e5ac9eaa8c1babd5d"} Dec 03 17:19:37 crc kubenswrapper[4758]: I1203 17:19:37.162525 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2878a24c-651b-46dc-87fd-78243a14aae0" path="/var/lib/kubelet/pods/2878a24c-651b-46dc-87fd-78243a14aae0/volumes" Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.063231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f4d5ff96-vfvv7" event={"ID":"d0887aea-e60c-4517-9c8c-c96498f912a4","Type":"ContainerStarted","Data":"0b6b8750178fa432f12fa07d0ae26570d538a6021da9fde9be6813c446d4d7ee"} Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.067085 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api-log" containerID="cri-o://a02e8862636694ec285582045f7961338abf66bf4ea94c2855817fa9605202b0" gracePeriod=30 Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.067366 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api" containerID="cri-o://c48ac1d12c26d1ca748d513e2f7db2a7890523ce01984d3f8fb15ee6354d35a0" gracePeriod=30 Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.066902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd65cf16-26a7-4dc2-b39e-23bc4afcb560","Type":"ContainerStarted","Data":"c48ac1d12c26d1ca748d513e2f7db2a7890523ce01984d3f8fb15ee6354d35a0"} Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.079204 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.102740 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.102698737 podStartE2EDuration="6.102698737s" podCreationTimestamp="2025-12-03 17:19:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:38.09196967 +0000 UTC m=+1433.293346531" watchObservedRunningTime="2025-12-03 17:19:38.102698737 +0000 UTC m=+1433.304075598" Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.128665 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerStarted","Data":"8595de5956503317e9c9965ad2288163bbbdf86eb190728548bc8ff6fb456660"} Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.136069 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"042060a5-b97e-49a0-9d5f-f946c785cdf3","Type":"ContainerStarted","Data":"0f334e972f3c44ece36746d94eaef7f1fd3b308b0f269d5c7b4b599e1ad57960"} Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.978557 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:38 crc kubenswrapper[4758]: I1203 17:19:38.982503 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.081021 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65c99cf6db-25flz"] Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.081343 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65c99cf6db-25flz" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api-log" containerID="cri-o://4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f" gracePeriod=30 Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.081953 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65c99cf6db-25flz" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api" containerID="cri-o://edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e" gracePeriod=30 Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.159729 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f4d5ff96-vfvv7" event={"ID":"d0887aea-e60c-4517-9c8c-c96498f912a4","Type":"ContainerStarted","Data":"34795b0ebc5422c2fc4ff2442f6fa1ac811ee787cb0ea6513b2b2de3fde194d6"} Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.177847 4758 generic.go:334] "Generic (PLEG): container finished" podID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerID="a02e8862636694ec285582045f7961338abf66bf4ea94c2855817fa9605202b0" exitCode=143 Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.177968 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd65cf16-26a7-4dc2-b39e-23bc4afcb560","Type":"ContainerDied","Data":"a02e8862636694ec285582045f7961338abf66bf4ea94c2855817fa9605202b0"} Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.195412 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" event={"ID":"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf","Type":"ContainerStarted","Data":"24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e"} Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.255489 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69dc74dcbf-wzmb5"] Dec 03 17:19:39 crc kubenswrapper[4758]: E1203 17:19:39.256069 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2878a24c-651b-46dc-87fd-78243a14aae0" containerName="init" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.256092 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2878a24c-651b-46dc-87fd-78243a14aae0" containerName="init" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.256329 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2878a24c-651b-46dc-87fd-78243a14aae0" containerName="init" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.257613 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.263576 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.263619 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.289654 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69dc74dcbf-wzmb5"] Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.392997 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-public-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.393155 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-internal-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.395371 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-ovndb-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.395466 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-httpd-config\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.395536 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-config\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.395692 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9vdz\" (UniqueName: \"kubernetes.io/projected/bd709657-fb8d-4c81-873f-a0b44d63abef-kube-api-access-c9vdz\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.395793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-combined-ca-bundle\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498299 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-public-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-internal-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498418 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-ovndb-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498441 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-httpd-config\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498464 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-config\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498495 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9vdz\" (UniqueName: \"kubernetes.io/projected/bd709657-fb8d-4c81-873f-a0b44d63abef-kube-api-access-c9vdz\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.498523 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-combined-ca-bundle\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.513052 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-ovndb-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.513455 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-internal-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.514340 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-config\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.536534 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-public-tls-certs\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.537121 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-httpd-config\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.537505 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-combined-ca-bundle\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.554518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9vdz\" (UniqueName: \"kubernetes.io/projected/bd709657-fb8d-4c81-873f-a0b44d63abef-kube-api-access-c9vdz\") pod \"neutron-69dc74dcbf-wzmb5\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:39 crc kubenswrapper[4758]: I1203 17:19:39.579262 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:40 crc kubenswrapper[4758]: I1203 17:19:40.273174 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-gxqxh" podUID="272d2f0a-3217-4e67-82f3-43dc7aa334f5" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:19:41 crc kubenswrapper[4758]: I1203 17:19:41.220908 4758 generic.go:334] "Generic (PLEG): container finished" podID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerID="4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f" exitCode=143 Dec 03 17:19:41 crc kubenswrapper[4758]: I1203 17:19:41.221299 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65c99cf6db-25flz" event={"ID":"b85bc93b-b75d-4d54-aa50-cc9027c0f044","Type":"ContainerDied","Data":"4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f"} Dec 03 17:19:41 crc kubenswrapper[4758]: I1203 17:19:41.223333 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"042060a5-b97e-49a0-9d5f-f946c785cdf3","Type":"ContainerStarted","Data":"cb32d3d98ed7e578711db4fcfe927af709cfe713f86240a09219099fc13c341f"} Dec 03 17:19:41 crc kubenswrapper[4758]: I1203 17:19:41.652710 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69dc74dcbf-wzmb5"] Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.248967 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dc74dcbf-wzmb5" event={"ID":"bd709657-fb8d-4c81-873f-a0b44d63abef","Type":"ContainerStarted","Data":"2496f74c6aa8a3c995f6fcb782d022aa7cb8ab73e8224330eecc746a01760d7c"} Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.249438 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dc74dcbf-wzmb5" event={"ID":"bd709657-fb8d-4c81-873f-a0b44d63abef","Type":"ContainerStarted","Data":"6083b0d716f3869aee6fc96831264050489b32214dcf21d56e722d859870aea3"} Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.253009 4758 generic.go:334] "Generic (PLEG): container finished" podID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerID="c48ac1d12c26d1ca748d513e2f7db2a7890523ce01984d3f8fb15ee6354d35a0" exitCode=0 Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.253063 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd65cf16-26a7-4dc2-b39e-23bc4afcb560","Type":"ContainerDied","Data":"c48ac1d12c26d1ca748d513e2f7db2a7890523ce01984d3f8fb15ee6354d35a0"} Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.256191 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerStarted","Data":"8c63a18309529436bd161ad90e515e1c0b282024a6c03715a18c56b19db6feb6"} Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.257163 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.288795 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" podStartSLOduration=9.288768154 podStartE2EDuration="9.288768154s" podCreationTimestamp="2025-12-03 17:19:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:42.28340558 +0000 UTC m=+1437.484782441" watchObservedRunningTime="2025-12-03 17:19:42.288768154 +0000 UTC m=+1437.490145015" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.314150 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.648606966 podStartE2EDuration="11.314124394s" podCreationTimestamp="2025-12-03 17:19:31 +0000 UTC" firstStartedPulling="2025-12-03 17:19:33.022320568 +0000 UTC m=+1428.223697419" lastFinishedPulling="2025-12-03 17:19:41.687837986 +0000 UTC m=+1436.889214847" observedRunningTime="2025-12-03 17:19:42.308811882 +0000 UTC m=+1437.510188743" watchObservedRunningTime="2025-12-03 17:19:42.314124394 +0000 UTC m=+1437.515501255" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.337308 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64f4d5ff96-vfvv7" podStartSLOduration=8.337282866 podStartE2EDuration="8.337282866s" podCreationTimestamp="2025-12-03 17:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:42.332046106 +0000 UTC m=+1437.533422977" watchObservedRunningTime="2025-12-03 17:19:42.337282866 +0000 UTC m=+1437.538659727" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.358123 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.360421 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.159:8080/\": dial tcp 10.217.0.159:8080: connect: connection refused" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.587695 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65c99cf6db-25flz" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:38538->10.217.0.156:9311: read: connection reset by peer" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.587782 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65c99cf6db-25flz" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:38548->10.217.0.156:9311: read: connection reset by peer" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.822940 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.872673 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=10.131963667 podStartE2EDuration="11.872649824s" podCreationTimestamp="2025-12-03 17:19:31 +0000 UTC" firstStartedPulling="2025-12-03 17:19:33.124614904 +0000 UTC m=+1428.325991765" lastFinishedPulling="2025-12-03 17:19:34.865301071 +0000 UTC m=+1430.066677922" observedRunningTime="2025-12-03 17:19:42.371892645 +0000 UTC m=+1437.573269526" watchObservedRunningTime="2025-12-03 17:19:42.872649824 +0000 UTC m=+1438.074026675" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.891589 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-logs\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.891726 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-combined-ca-bundle\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.891775 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data-custom\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.891802 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.891912 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-scripts\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.891967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgfwm\" (UniqueName: \"kubernetes.io/projected/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-kube-api-access-dgfwm\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.892020 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-etc-machine-id\") pod \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\" (UID: \"fd65cf16-26a7-4dc2-b39e-23bc4afcb560\") " Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.892562 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.893924 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-logs" (OuterVolumeSpecName: "logs") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.899928 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-scripts" (OuterVolumeSpecName: "scripts") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.899956 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.902957 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-kube-api-access-dgfwm" (OuterVolumeSpecName: "kube-api-access-dgfwm") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "kube-api-access-dgfwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.953186 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.994888 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.994922 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgfwm\" (UniqueName: \"kubernetes.io/projected/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-kube-api-access-dgfwm\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.994938 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.994948 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.994957 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:42 crc kubenswrapper[4758]: I1203 17:19:42.994967 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.014554 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data" (OuterVolumeSpecName: "config-data") pod "fd65cf16-26a7-4dc2-b39e-23bc4afcb560" (UID: "fd65cf16-26a7-4dc2-b39e-23bc4afcb560"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.097249 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65cf16-26a7-4dc2-b39e-23bc4afcb560-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.196372 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.290364 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dc74dcbf-wzmb5" event={"ID":"bd709657-fb8d-4c81-873f-a0b44d63abef","Type":"ContainerStarted","Data":"6637a67a10c678e5c8c666412014de74c5391a41c691d978da31902186b3e7c0"} Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.292144 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.301954 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data-custom\") pod \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.302017 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data\") pod \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.302198 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t82rf\" (UniqueName: \"kubernetes.io/projected/b85bc93b-b75d-4d54-aa50-cc9027c0f044-kube-api-access-t82rf\") pod \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.302288 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85bc93b-b75d-4d54-aa50-cc9027c0f044-logs\") pod \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.302401 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-combined-ca-bundle\") pod \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\" (UID: \"b85bc93b-b75d-4d54-aa50-cc9027c0f044\") " Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.303598 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd65cf16-26a7-4dc2-b39e-23bc4afcb560","Type":"ContainerDied","Data":"2b2d1235f2864db4211d0d34d838174306fb60db187db0a533cbd729a18d35e1"} Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.303691 4758 scope.go:117] "RemoveContainer" containerID="c48ac1d12c26d1ca748d513e2f7db2a7890523ce01984d3f8fb15ee6354d35a0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.303933 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.307541 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b85bc93b-b75d-4d54-aa50-cc9027c0f044-logs" (OuterVolumeSpecName: "logs") pod "b85bc93b-b75d-4d54-aa50-cc9027c0f044" (UID: "b85bc93b-b75d-4d54-aa50-cc9027c0f044"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.310509 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85bc93b-b75d-4d54-aa50-cc9027c0f044-kube-api-access-t82rf" (OuterVolumeSpecName: "kube-api-access-t82rf") pod "b85bc93b-b75d-4d54-aa50-cc9027c0f044" (UID: "b85bc93b-b75d-4d54-aa50-cc9027c0f044"). InnerVolumeSpecName "kube-api-access-t82rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.315142 4758 generic.go:334] "Generic (PLEG): container finished" podID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerID="edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e" exitCode=0 Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.315386 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65c99cf6db-25flz" event={"ID":"b85bc93b-b75d-4d54-aa50-cc9027c0f044","Type":"ContainerDied","Data":"edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e"} Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.324139 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65c99cf6db-25flz" event={"ID":"b85bc93b-b75d-4d54-aa50-cc9027c0f044","Type":"ContainerDied","Data":"b5f64dc80b3325a499d2176db7720bba2ebd55e5cb0f3d970ab5965aed52954b"} Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.315456 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65c99cf6db-25flz" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.326296 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.335077 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b85bc93b-b75d-4d54-aa50-cc9027c0f044" (UID: "b85bc93b-b75d-4d54-aa50-cc9027c0f044"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.335915 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.347237 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69dc74dcbf-wzmb5" podStartSLOduration=4.34717148 podStartE2EDuration="4.34717148s" podCreationTimestamp="2025-12-03 17:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:43.326991368 +0000 UTC m=+1438.528368239" watchObservedRunningTime="2025-12-03 17:19:43.34717148 +0000 UTC m=+1438.548548361" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.383823 4758 scope.go:117] "RemoveContainer" containerID="a02e8862636694ec285582045f7961338abf66bf4ea94c2855817fa9605202b0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.408041 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t82rf\" (UniqueName: \"kubernetes.io/projected/b85bc93b-b75d-4d54-aa50-cc9027c0f044-kube-api-access-t82rf\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.408091 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85bc93b-b75d-4d54-aa50-cc9027c0f044-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.408107 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.423652 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b85bc93b-b75d-4d54-aa50-cc9027c0f044" (UID: "b85bc93b-b75d-4d54-aa50-cc9027c0f044"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.447213 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.455785 4758 scope.go:117] "RemoveContainer" containerID="edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.462297 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.480572 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data" (OuterVolumeSpecName: "config-data") pod "b85bc93b-b75d-4d54-aa50-cc9027c0f044" (UID: "b85bc93b-b75d-4d54-aa50-cc9027c0f044"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.495762 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:43 crc kubenswrapper[4758]: E1203 17:19:43.496422 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.496449 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api" Dec 03 17:19:43 crc kubenswrapper[4758]: E1203 17:19:43.496471 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.496479 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api" Dec 03 17:19:43 crc kubenswrapper[4758]: E1203 17:19:43.496502 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api-log" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.496511 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api-log" Dec 03 17:19:43 crc kubenswrapper[4758]: E1203 17:19:43.496543 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api-log" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.496550 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api-log" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.496945 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.497001 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.497020 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" containerName="barbican-api-log" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.497033 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" containerName="cinder-api-log" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.498450 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.505774 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.506148 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.506292 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.511664 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.511718 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b85bc93b-b75d-4d54-aa50-cc9027c0f044-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.527696 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bxf5h"] Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.528019 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerName="dnsmasq-dns" containerID="cri-o://25065c61a15ddba760fbd565f929949c6cd3b972d95f83ea74bea465a8ba22dc" gracePeriod=10 Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.539897 4758 scope.go:117] "RemoveContainer" containerID="4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.548544 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.613719 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-scripts\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.614242 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.614472 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.614634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7plrd\" (UniqueName: \"kubernetes.io/projected/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-kube-api-access-7plrd\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.614804 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.614956 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.615158 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.615420 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data-custom\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.615551 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-logs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.667022 4758 scope.go:117] "RemoveContainer" containerID="edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e" Dec 03 17:19:43 crc kubenswrapper[4758]: E1203 17:19:43.672955 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e\": container with ID starting with edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e not found: ID does not exist" containerID="edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.673338 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e"} err="failed to get container status \"edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e\": rpc error: code = NotFound desc = could not find container \"edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e\": container with ID starting with edce707962aa509859d84658f436298835d9ebe5ed3f1673b4bfcb0d1e11ff2e not found: ID does not exist" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.673494 4758 scope.go:117] "RemoveContainer" containerID="4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f" Dec 03 17:19:43 crc kubenswrapper[4758]: E1203 17:19:43.690014 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f\": container with ID starting with 4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f not found: ID does not exist" containerID="4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.690570 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f"} err="failed to get container status \"4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f\": rpc error: code = NotFound desc = could not find container \"4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f\": container with ID starting with 4e5ade69bbee00a57a7e9d74c7d94bb9080ad04d0d10430e1cd1d02cfcad106f not found: ID does not exist" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.736492 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.736721 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.736826 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7plrd\" (UniqueName: \"kubernetes.io/projected/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-kube-api-access-7plrd\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.736921 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.736972 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.737086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.737329 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data-custom\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.737377 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-logs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.737406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-scripts\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.743914 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.744389 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-logs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.748814 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-scripts\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.749976 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.751069 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.756558 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.757155 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data-custom\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.761869 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.778789 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7plrd\" (UniqueName: \"kubernetes.io/projected/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-kube-api-access-7plrd\") pod \"cinder-api-0\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " pod="openstack/cinder-api-0" Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.778900 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65c99cf6db-25flz"] Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.835061 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-65c99cf6db-25flz"] Dec 03 17:19:43 crc kubenswrapper[4758]: I1203 17:19:43.858353 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.378555 4758 generic.go:334] "Generic (PLEG): container finished" podID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerID="25065c61a15ddba760fbd565f929949c6cd3b972d95f83ea74bea465a8ba22dc" exitCode=0 Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.380039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" event={"ID":"4401ac8b-420e-41b9-a6b0-aba2886b8ff8","Type":"ContainerDied","Data":"25065c61a15ddba760fbd565f929949c6cd3b972d95f83ea74bea465a8ba22dc"} Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.380088 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" event={"ID":"4401ac8b-420e-41b9-a6b0-aba2886b8ff8","Type":"ContainerDied","Data":"cc1b52602f084ab721084d07659a360a9a4947eadf99b3c98f294744a9a1a01d"} Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.380104 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc1b52602f084ab721084d07659a360a9a4947eadf99b3c98f294744a9a1a01d" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.400186 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.474473 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-swift-storage-0\") pod \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.474556 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-sb\") pod \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.474704 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-config\") pod \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.474774 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-nb\") pod \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.474813 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-svc\") pod \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.474850 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kks7s\" (UniqueName: \"kubernetes.io/projected/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-kube-api-access-kks7s\") pod \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\" (UID: \"4401ac8b-420e-41b9-a6b0-aba2886b8ff8\") " Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.509944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-kube-api-access-kks7s" (OuterVolumeSpecName: "kube-api-access-kks7s") pod "4401ac8b-420e-41b9-a6b0-aba2886b8ff8" (UID: "4401ac8b-420e-41b9-a6b0-aba2886b8ff8"). InnerVolumeSpecName "kube-api-access-kks7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.539815 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.557480 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4401ac8b-420e-41b9-a6b0-aba2886b8ff8" (UID: "4401ac8b-420e-41b9-a6b0-aba2886b8ff8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.566503 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.567639 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.581511 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.581955 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kks7s\" (UniqueName: \"kubernetes.io/projected/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-kube-api-access-kks7s\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.599580 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4401ac8b-420e-41b9-a6b0-aba2886b8ff8" (UID: "4401ac8b-420e-41b9-a6b0-aba2886b8ff8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.622374 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-config" (OuterVolumeSpecName: "config") pod "4401ac8b-420e-41b9-a6b0-aba2886b8ff8" (UID: "4401ac8b-420e-41b9-a6b0-aba2886b8ff8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.635298 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4401ac8b-420e-41b9-a6b0-aba2886b8ff8" (UID: "4401ac8b-420e-41b9-a6b0-aba2886b8ff8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.679320 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4401ac8b-420e-41b9-a6b0-aba2886b8ff8" (UID: "4401ac8b-420e-41b9-a6b0-aba2886b8ff8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.684427 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.684467 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.684480 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:44 crc kubenswrapper[4758]: I1203 17:19:44.684491 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4401ac8b-420e-41b9-a6b0-aba2886b8ff8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:45 crc kubenswrapper[4758]: I1203 17:19:45.139058 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85bc93b-b75d-4d54-aa50-cc9027c0f044" path="/var/lib/kubelet/pods/b85bc93b-b75d-4d54-aa50-cc9027c0f044/volumes" Dec 03 17:19:45 crc kubenswrapper[4758]: I1203 17:19:45.148037 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd65cf16-26a7-4dc2-b39e-23bc4afcb560" path="/var/lib/kubelet/pods/fd65cf16-26a7-4dc2-b39e-23bc4afcb560/volumes" Dec 03 17:19:45 crc kubenswrapper[4758]: I1203 17:19:45.398742 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bxf5h" Dec 03 17:19:45 crc kubenswrapper[4758]: I1203 17:19:45.400300 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b","Type":"ContainerStarted","Data":"76de1c5f74c1f80efeadc927dff83e66fb87f4001807e5daeeafe7f4f580cfc3"} Dec 03 17:19:45 crc kubenswrapper[4758]: I1203 17:19:45.526757 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bxf5h"] Dec 03 17:19:45 crc kubenswrapper[4758]: I1203 17:19:45.546625 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bxf5h"] Dec 03 17:19:46 crc kubenswrapper[4758]: I1203 17:19:46.455288 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b","Type":"ContainerStarted","Data":"b515931c8d15bc7880738113558415685d689f0ef349e841890a0dc9c902625e"} Dec 03 17:19:46 crc kubenswrapper[4758]: I1203 17:19:46.787898 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:19:47 crc kubenswrapper[4758]: I1203 17:19:47.133468 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" path="/var/lib/kubelet/pods/4401ac8b-420e-41b9-a6b0-aba2886b8ff8/volumes" Dec 03 17:19:47 crc kubenswrapper[4758]: I1203 17:19:47.496909 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b","Type":"ContainerStarted","Data":"ef163a1a1695789c803e3b2391ac85959aee35c4dc735cf405bb6587fe19ed5f"} Dec 03 17:19:47 crc kubenswrapper[4758]: I1203 17:19:47.498323 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 17:19:47 crc kubenswrapper[4758]: I1203 17:19:47.547446 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.5474183870000005 podStartE2EDuration="4.547418387s" podCreationTimestamp="2025-12-03 17:19:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:47.535125697 +0000 UTC m=+1442.736502558" watchObservedRunningTime="2025-12-03 17:19:47.547418387 +0000 UTC m=+1442.748795248" Dec 03 17:19:47 crc kubenswrapper[4758]: I1203 17:19:47.827621 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 17:19:47 crc kubenswrapper[4758]: I1203 17:19:47.902321 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.477441 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:48 crc kubenswrapper[4758]: E1203 17:19:48.479123 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerName="dnsmasq-dns" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.479149 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerName="dnsmasq-dns" Dec 03 17:19:48 crc kubenswrapper[4758]: E1203 17:19:48.479175 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerName="init" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.479184 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerName="init" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.479393 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4401ac8b-420e-41b9-a6b0-aba2886b8ff8" containerName="dnsmasq-dns" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.480149 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.483011 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.485269 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-spmkg" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.485474 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.505828 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.534339 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="cinder-scheduler" containerID="cri-o://0f334e972f3c44ece36746d94eaef7f1fd3b308b0f269d5c7b4b599e1ad57960" gracePeriod=30 Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.534788 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="probe" containerID="cri-o://cb32d3d98ed7e578711db4fcfe927af709cfe713f86240a09219099fc13c341f" gracePeriod=30 Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.612251 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.612316 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config-secret\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.612926 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snntw\" (UniqueName: \"kubernetes.io/projected/8929c72b-b422-4dd1-913d-434d8ea93817-kube-api-access-snntw\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.613080 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.715254 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snntw\" (UniqueName: \"kubernetes.io/projected/8929c72b-b422-4dd1-913d-434d8ea93817-kube-api-access-snntw\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.715328 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.715422 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.715440 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config-secret\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.717982 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.723737 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.742224 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config-secret\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.742968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snntw\" (UniqueName: \"kubernetes.io/projected/8929c72b-b422-4dd1-913d-434d8ea93817-kube-api-access-snntw\") pod \"openstackclient\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.771302 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.772622 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.808651 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.839774 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.841283 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:48 crc kubenswrapper[4758]: I1203 17:19:48.849769 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:48 crc kubenswrapper[4758]: E1203 17:19:48.940966 4758 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 17:19:48 crc kubenswrapper[4758]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_8929c72b-b422-4dd1-913d-434d8ea93817_0(2c34d5688dac671a39af355301d65f384fa4e4004efa74ce1f62e20c34d27b51): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2c34d5688dac671a39af355301d65f384fa4e4004efa74ce1f62e20c34d27b51" Netns:"/var/run/netns/29127544-6099-47a8-a835-1eba86dae6f4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=2c34d5688dac671a39af355301d65f384fa4e4004efa74ce1f62e20c34d27b51;K8S_POD_UID=8929c72b-b422-4dd1-913d-434d8ea93817" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/8929c72b-b422-4dd1-913d-434d8ea93817]: expected pod UID "8929c72b-b422-4dd1-913d-434d8ea93817" but got "790de883-aaa9-4faa-b18c-61ef4d093ab4" from Kube API Dec 03 17:19:48 crc kubenswrapper[4758]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 17:19:48 crc kubenswrapper[4758]: > Dec 03 17:19:48 crc kubenswrapper[4758]: E1203 17:19:48.941128 4758 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 17:19:48 crc kubenswrapper[4758]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_8929c72b-b422-4dd1-913d-434d8ea93817_0(2c34d5688dac671a39af355301d65f384fa4e4004efa74ce1f62e20c34d27b51): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2c34d5688dac671a39af355301d65f384fa4e4004efa74ce1f62e20c34d27b51" Netns:"/var/run/netns/29127544-6099-47a8-a835-1eba86dae6f4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=2c34d5688dac671a39af355301d65f384fa4e4004efa74ce1f62e20c34d27b51;K8S_POD_UID=8929c72b-b422-4dd1-913d-434d8ea93817" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/8929c72b-b422-4dd1-913d-434d8ea93817]: expected pod UID "8929c72b-b422-4dd1-913d-434d8ea93817" but got "790de883-aaa9-4faa-b18c-61ef4d093ab4" from Kube API Dec 03 17:19:48 crc kubenswrapper[4758]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 17:19:48 crc kubenswrapper[4758]: > pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.021977 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.022098 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config-secret\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.022281 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nfkk\" (UniqueName: \"kubernetes.io/projected/790de883-aaa9-4faa-b18c-61ef4d093ab4-kube-api-access-9nfkk\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.022326 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.125775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.127217 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.131367 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config-secret\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.131790 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nfkk\" (UniqueName: \"kubernetes.io/projected/790de883-aaa9-4faa-b18c-61ef4d093ab4-kube-api-access-9nfkk\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.131839 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.138949 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.146652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config-secret\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.156928 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nfkk\" (UniqueName: \"kubernetes.io/projected/790de883-aaa9-4faa-b18c-61ef4d093ab4-kube-api-access-9nfkk\") pod \"openstackclient\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.298913 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.552274 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.563908 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.568012 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8929c72b-b422-4dd1-913d-434d8ea93817" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.645095 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-combined-ca-bundle\") pod \"8929c72b-b422-4dd1-913d-434d8ea93817\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.645182 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snntw\" (UniqueName: \"kubernetes.io/projected/8929c72b-b422-4dd1-913d-434d8ea93817-kube-api-access-snntw\") pod \"8929c72b-b422-4dd1-913d-434d8ea93817\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.645373 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config-secret\") pod \"8929c72b-b422-4dd1-913d-434d8ea93817\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.645472 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config\") pod \"8929c72b-b422-4dd1-913d-434d8ea93817\" (UID: \"8929c72b-b422-4dd1-913d-434d8ea93817\") " Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.646143 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8929c72b-b422-4dd1-913d-434d8ea93817" (UID: "8929c72b-b422-4dd1-913d-434d8ea93817"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.651806 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8929c72b-b422-4dd1-913d-434d8ea93817" (UID: "8929c72b-b422-4dd1-913d-434d8ea93817"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.651955 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8929c72b-b422-4dd1-913d-434d8ea93817" (UID: "8929c72b-b422-4dd1-913d-434d8ea93817"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.651978 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8929c72b-b422-4dd1-913d-434d8ea93817-kube-api-access-snntw" (OuterVolumeSpecName: "kube-api-access-snntw") pod "8929c72b-b422-4dd1-913d-434d8ea93817" (UID: "8929c72b-b422-4dd1-913d-434d8ea93817"). InnerVolumeSpecName "kube-api-access-snntw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.747236 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.747283 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snntw\" (UniqueName: \"kubernetes.io/projected/8929c72b-b422-4dd1-913d-434d8ea93817-kube-api-access-snntw\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.747298 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.747307 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8929c72b-b422-4dd1-913d-434d8ea93817-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:49 crc kubenswrapper[4758]: I1203 17:19:49.801084 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 17:19:50 crc kubenswrapper[4758]: I1203 17:19:50.571036 4758 generic.go:334] "Generic (PLEG): container finished" podID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerID="cb32d3d98ed7e578711db4fcfe927af709cfe713f86240a09219099fc13c341f" exitCode=0 Dec 03 17:19:50 crc kubenswrapper[4758]: I1203 17:19:50.571113 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"042060a5-b97e-49a0-9d5f-f946c785cdf3","Type":"ContainerDied","Data":"cb32d3d98ed7e578711db4fcfe927af709cfe713f86240a09219099fc13c341f"} Dec 03 17:19:50 crc kubenswrapper[4758]: I1203 17:19:50.574456 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:19:50 crc kubenswrapper[4758]: I1203 17:19:50.574763 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"790de883-aaa9-4faa-b18c-61ef4d093ab4","Type":"ContainerStarted","Data":"c1360a0b66c06fc2eb9fdbe369533735c7d1905d09a6fa35616701793c466fca"} Dec 03 17:19:50 crc kubenswrapper[4758]: I1203 17:19:50.599343 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8929c72b-b422-4dd1-913d-434d8ea93817" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.129671 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8929c72b-b422-4dd1-913d-434d8ea93817" path="/var/lib/kubelet/pods/8929c72b-b422-4dd1-913d-434d8ea93817/volumes" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.606212 4758 generic.go:334] "Generic (PLEG): container finished" podID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerID="0f334e972f3c44ece36746d94eaef7f1fd3b308b0f269d5c7b4b599e1ad57960" exitCode=0 Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.606732 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"042060a5-b97e-49a0-9d5f-f946c785cdf3","Type":"ContainerDied","Data":"0f334e972f3c44ece36746d94eaef7f1fd3b308b0f269d5c7b4b599e1ad57960"} Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.859746 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.903247 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-scripts\") pod \"042060a5-b97e-49a0-9d5f-f946c785cdf3\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.903333 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data\") pod \"042060a5-b97e-49a0-9d5f-f946c785cdf3\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.903382 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data-custom\") pod \"042060a5-b97e-49a0-9d5f-f946c785cdf3\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.903505 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/042060a5-b97e-49a0-9d5f-f946c785cdf3-etc-machine-id\") pod \"042060a5-b97e-49a0-9d5f-f946c785cdf3\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.903536 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-combined-ca-bundle\") pod \"042060a5-b97e-49a0-9d5f-f946c785cdf3\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.903578 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kmd8\" (UniqueName: \"kubernetes.io/projected/042060a5-b97e-49a0-9d5f-f946c785cdf3-kube-api-access-9kmd8\") pod \"042060a5-b97e-49a0-9d5f-f946c785cdf3\" (UID: \"042060a5-b97e-49a0-9d5f-f946c785cdf3\") " Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.908790 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/042060a5-b97e-49a0-9d5f-f946c785cdf3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "042060a5-b97e-49a0-9d5f-f946c785cdf3" (UID: "042060a5-b97e-49a0-9d5f-f946c785cdf3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.920771 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/042060a5-b97e-49a0-9d5f-f946c785cdf3-kube-api-access-9kmd8" (OuterVolumeSpecName: "kube-api-access-9kmd8") pod "042060a5-b97e-49a0-9d5f-f946c785cdf3" (UID: "042060a5-b97e-49a0-9d5f-f946c785cdf3"). InnerVolumeSpecName "kube-api-access-9kmd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.925459 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-scripts" (OuterVolumeSpecName: "scripts") pod "042060a5-b97e-49a0-9d5f-f946c785cdf3" (UID: "042060a5-b97e-49a0-9d5f-f946c785cdf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.944456 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "042060a5-b97e-49a0-9d5f-f946c785cdf3" (UID: "042060a5-b97e-49a0-9d5f-f946c785cdf3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:51 crc kubenswrapper[4758]: I1203 17:19:51.992048 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "042060a5-b97e-49a0-9d5f-f946c785cdf3" (UID: "042060a5-b97e-49a0-9d5f-f946c785cdf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.008689 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kmd8\" (UniqueName: \"kubernetes.io/projected/042060a5-b97e-49a0-9d5f-f946c785cdf3-kube-api-access-9kmd8\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.008731 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.008756 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.008764 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/042060a5-b97e-49a0-9d5f-f946c785cdf3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.008775 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.134895 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data" (OuterVolumeSpecName: "config-data") pod "042060a5-b97e-49a0-9d5f-f946c785cdf3" (UID: "042060a5-b97e-49a0-9d5f-f946c785cdf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.214411 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042060a5-b97e-49a0-9d5f-f946c785cdf3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.621572 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"042060a5-b97e-49a0-9d5f-f946c785cdf3","Type":"ContainerDied","Data":"d2942bfce12e33f4021f5395fed3366141da6a143cf6af0bdaf0b1ffa20dc224"} Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.621639 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.622015 4758 scope.go:117] "RemoveContainer" containerID="cb32d3d98ed7e578711db4fcfe927af709cfe713f86240a09219099fc13c341f" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.667781 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.694102 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.707886 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.708100 4758 scope.go:117] "RemoveContainer" containerID="0f334e972f3c44ece36746d94eaef7f1fd3b308b0f269d5c7b4b599e1ad57960" Dec 03 17:19:52 crc kubenswrapper[4758]: E1203 17:19:52.708428 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="probe" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.708457 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="probe" Dec 03 17:19:52 crc kubenswrapper[4758]: E1203 17:19:52.708472 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="cinder-scheduler" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.708479 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="cinder-scheduler" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.708938 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="probe" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.708965 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" containerName="cinder-scheduler" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.710155 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.713965 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.728068 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.826485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.826559 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.826594 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.826625 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-scripts\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.826785 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/deb8eeb7-c8f9-4324-a518-b837c3ad043c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.826822 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2tth\" (UniqueName: \"kubernetes.io/projected/deb8eeb7-c8f9-4324-a518-b837c3ad043c-kube-api-access-w2tth\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928272 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/deb8eeb7-c8f9-4324-a518-b837c3ad043c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928338 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2tth\" (UniqueName: \"kubernetes.io/projected/deb8eeb7-c8f9-4324-a518-b837c3ad043c-kube-api-access-w2tth\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928400 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928441 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928473 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/deb8eeb7-c8f9-4324-a518-b837c3ad043c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.928504 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-scripts\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.934833 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-scripts\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.935025 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.937646 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.939158 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:52 crc kubenswrapper[4758]: I1203 17:19:52.952720 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2tth\" (UniqueName: \"kubernetes.io/projected/deb8eeb7-c8f9-4324-a518-b837c3ad043c-kube-api-access-w2tth\") pod \"cinder-scheduler-0\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " pod="openstack/cinder-scheduler-0" Dec 03 17:19:53 crc kubenswrapper[4758]: I1203 17:19:53.048472 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:19:53 crc kubenswrapper[4758]: I1203 17:19:53.136979 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="042060a5-b97e-49a0-9d5f-f946c785cdf3" path="/var/lib/kubelet/pods/042060a5-b97e-49a0-9d5f-f946c785cdf3/volumes" Dec 03 17:19:53 crc kubenswrapper[4758]: I1203 17:19:53.583485 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:19:53 crc kubenswrapper[4758]: W1203 17:19:53.598921 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeb8eeb7_c8f9_4324_a518_b837c3ad043c.slice/crio-f525490acf63bb49edc7ef1a5ca45e4acb7b684cf0db3b5fc21956513876cd11 WatchSource:0}: Error finding container f525490acf63bb49edc7ef1a5ca45e4acb7b684cf0db3b5fc21956513876cd11: Status 404 returned error can't find the container with id f525490acf63bb49edc7ef1a5ca45e4acb7b684cf0db3b5fc21956513876cd11 Dec 03 17:19:53 crc kubenswrapper[4758]: I1203 17:19:53.639557 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"deb8eeb7-c8f9-4324-a518-b837c3ad043c","Type":"ContainerStarted","Data":"f525490acf63bb49edc7ef1a5ca45e4acb7b684cf0db3b5fc21956513876cd11"} Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.057189 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.057864 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-central-agent" containerID="cri-o://d4fdb1467bc7e1d7119c3cd23b4a6f9c252938f343e72f940f4800aed1ad0cd9" gracePeriod=30 Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.058265 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="sg-core" containerID="cri-o://8595de5956503317e9c9965ad2288163bbbdf86eb190728548bc8ff6fb456660" gracePeriod=30 Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.058533 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="proxy-httpd" containerID="cri-o://8c63a18309529436bd161ad90e515e1c0b282024a6c03715a18c56b19db6feb6" gracePeriod=30 Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.058542 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-notification-agent" containerID="cri-o://9299b9e27e5198daa91a90f721e702d0ea26aa93e94e1904454efd0443ccf678" gracePeriod=30 Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.110642 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": read tcp 10.217.0.2:49926->10.217.0.158:3000: read: connection reset by peer" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.599805 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6fc8f59f6f-7zxl4"] Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.602327 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.605391 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.605655 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.605913 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.608839 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6fc8f59f6f-7zxl4"] Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682122 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-combined-ca-bundle\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682182 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-run-httpd\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682239 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-internal-tls-certs\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682285 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-config-data\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682304 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlbv8\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-kube-api-access-qlbv8\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682379 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-etc-swift\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682416 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-log-httpd\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.682434 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-public-tls-certs\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.737357 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"deb8eeb7-c8f9-4324-a518-b837c3ad043c","Type":"ContainerStarted","Data":"df8bc0faf1cabc23c62d9aea7423ca778a9372ef7f9e311140e90eae1b53b014"} Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.751187 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerID="8c63a18309529436bd161ad90e515e1c0b282024a6c03715a18c56b19db6feb6" exitCode=0 Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.751241 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerID="8595de5956503317e9c9965ad2288163bbbdf86eb190728548bc8ff6fb456660" exitCode=2 Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.751268 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerDied","Data":"8c63a18309529436bd161ad90e515e1c0b282024a6c03715a18c56b19db6feb6"} Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.751306 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerDied","Data":"8595de5956503317e9c9965ad2288163bbbdf86eb190728548bc8ff6fb456660"} Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785250 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-etc-swift\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785355 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-log-httpd\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785379 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-public-tls-certs\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785424 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-combined-ca-bundle\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785449 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-run-httpd\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785494 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-internal-tls-certs\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785541 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-config-data\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.785557 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlbv8\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-kube-api-access-qlbv8\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.787583 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-log-httpd\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.788349 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-run-httpd\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.794941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-internal-tls-certs\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.803135 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-etc-swift\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.809298 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-config-data\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.810549 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-public-tls-certs\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.812593 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-combined-ca-bundle\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.818314 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlbv8\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-kube-api-access-qlbv8\") pod \"swift-proxy-6fc8f59f6f-7zxl4\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:54 crc kubenswrapper[4758]: I1203 17:19:54.955384 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:19:55 crc kubenswrapper[4758]: I1203 17:19:55.772125 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"deb8eeb7-c8f9-4324-a518-b837c3ad043c","Type":"ContainerStarted","Data":"a8202adddc1c7cf7ab92d5ed5e61b0495c0d800596e0b0472d95b872e5eca55f"} Dec 03 17:19:55 crc kubenswrapper[4758]: I1203 17:19:55.799093 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerID="d4fdb1467bc7e1d7119c3cd23b4a6f9c252938f343e72f940f4800aed1ad0cd9" exitCode=0 Dec 03 17:19:55 crc kubenswrapper[4758]: I1203 17:19:55.799515 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerDied","Data":"d4fdb1467bc7e1d7119c3cd23b4a6f9c252938f343e72f940f4800aed1ad0cd9"} Dec 03 17:19:55 crc kubenswrapper[4758]: I1203 17:19:55.806031 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.805986691 podStartE2EDuration="3.805986691s" podCreationTimestamp="2025-12-03 17:19:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:55.800920026 +0000 UTC m=+1451.002296887" watchObservedRunningTime="2025-12-03 17:19:55.805986691 +0000 UTC m=+1451.007363552" Dec 03 17:19:55 crc kubenswrapper[4758]: I1203 17:19:55.889376 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6fc8f59f6f-7zxl4"] Dec 03 17:19:56 crc kubenswrapper[4758]: I1203 17:19:56.960671 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 17:19:58 crc kubenswrapper[4758]: I1203 17:19:58.049306 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 17:19:58 crc kubenswrapper[4758]: I1203 17:19:58.852018 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerID="9299b9e27e5198daa91a90f721e702d0ea26aa93e94e1904454efd0443ccf678" exitCode=0 Dec 03 17:19:58 crc kubenswrapper[4758]: I1203 17:19:58.852513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerDied","Data":"9299b9e27e5198daa91a90f721e702d0ea26aa93e94e1904454efd0443ccf678"} Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.083823 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": dial tcp 10.217.0.158:3000: connect: connection refused" Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.730783 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.733290 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-log" containerID="cri-o://55c59d419d0dddb9e20a6a9054a34c3b529a7c39deb72363a316ff284231e1fe" gracePeriod=30 Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.733761 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-httpd" containerID="cri-o://ea6dcf9c2fc3a2c5e5cd9cbbf22a997692e23bc0e78284fb1d6304f44a979e57" gracePeriod=30 Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.970001 4758 generic.go:334] "Generic (PLEG): container finished" podID="74ca64c8-28c7-472e-85bd-acd6af307580" containerID="55c59d419d0dddb9e20a6a9054a34c3b529a7c39deb72363a316ff284231e1fe" exitCode=143 Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.970070 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ca64c8-28c7-472e-85bd-acd6af307580","Type":"ContainerDied","Data":"55c59d419d0dddb9e20a6a9054a34c3b529a7c39deb72363a316ff284231e1fe"} Dec 03 17:20:02 crc kubenswrapper[4758]: I1203 17:20:02.971995 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" event={"ID":"583229ad-ae2e-41b5-aad3-3c196b829f40","Type":"ContainerStarted","Data":"944f9d6d0e211ceb4b3ce57796bfff3ea944d6469168a780a09799cfa3845c76"} Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.019422 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.074120 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gphk4\" (UniqueName: \"kubernetes.io/projected/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-kube-api-access-gphk4\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.075079 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-run-httpd\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.075227 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-combined-ca-bundle\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.075738 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-config-data\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.075935 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.076165 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-scripts\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.076232 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-log-httpd\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.076268 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-sg-core-conf-yaml\") pod \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\" (UID: \"f3ae7804-bb04-49dd-844c-78bd4de0b1cb\") " Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.077258 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.078025 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.083251 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-kube-api-access-gphk4" (OuterVolumeSpecName: "kube-api-access-gphk4") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "kube-api-access-gphk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.084116 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-scripts" (OuterVolumeSpecName: "scripts") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.162089 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.180350 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.180399 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.180411 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.180422 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gphk4\" (UniqueName: \"kubernetes.io/projected/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-kube-api-access-gphk4\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.235233 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.282887 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.285495 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-config-data" (OuterVolumeSpecName: "config-data") pod "f3ae7804-bb04-49dd-844c-78bd4de0b1cb" (UID: "f3ae7804-bb04-49dd-844c-78bd4de0b1cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.381384 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.385885 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ae7804-bb04-49dd-844c-78bd4de0b1cb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.869960 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-d4zrc"] Dec 03 17:20:03 crc kubenswrapper[4758]: E1203 17:20:03.870913 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="sg-core" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.870936 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="sg-core" Dec 03 17:20:03 crc kubenswrapper[4758]: E1203 17:20:03.870963 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-notification-agent" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.870970 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-notification-agent" Dec 03 17:20:03 crc kubenswrapper[4758]: E1203 17:20:03.870977 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="proxy-httpd" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.870983 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="proxy-httpd" Dec 03 17:20:03 crc kubenswrapper[4758]: E1203 17:20:03.871003 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-central-agent" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.871010 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-central-agent" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.871271 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="sg-core" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.871308 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="proxy-httpd" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.871325 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-central-agent" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.871341 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" containerName="ceilometer-notification-agent" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.872366 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.886905 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-d4zrc"] Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.993655 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-frhtr"] Dec 03 17:20:03 crc kubenswrapper[4758]: I1203 17:20:03.995614 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.002620 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l8wx\" (UniqueName: \"kubernetes.io/projected/7ff4c92a-13c5-4a18-944d-38682d75ec07-kube-api-access-7l8wx\") pod \"nova-api-db-create-d4zrc\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.002740 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff4c92a-13c5-4a18-944d-38682d75ec07-operator-scripts\") pod \"nova-api-db-create-d4zrc\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.004389 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"790de883-aaa9-4faa-b18c-61ef4d093ab4","Type":"ContainerStarted","Data":"5659d420ac31a290e23ff03a8dee2a674a17edcf17f81a32fdb9e77fc73d7428"} Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.008631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" event={"ID":"583229ad-ae2e-41b5-aad3-3c196b829f40","Type":"ContainerStarted","Data":"6b35ae8e6d0ff4c3cd07edce5d23b46583d8b4af2ecf93f7e9477313990aa41c"} Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.008729 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" event={"ID":"583229ad-ae2e-41b5-aad3-3c196b829f40","Type":"ContainerStarted","Data":"033147cfa4974a6e9865e1dce4a3ab80379ee25004c515e9a54cca5bc2a76e0e"} Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.009054 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.009109 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.017239 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3ae7804-bb04-49dd-844c-78bd4de0b1cb","Type":"ContainerDied","Data":"a393b197db8b6dfd4f7659361896c2477cdabc5dc2fe07f41dc02bddc2c7477a"} Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.017321 4758 scope.go:117] "RemoveContainer" containerID="8c63a18309529436bd161ad90e515e1c0b282024a6c03715a18c56b19db6feb6" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.017370 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.031091 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-frhtr"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.054933 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c90a-account-create-update-24496"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.057733 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.058284 4758 scope.go:117] "RemoveContainer" containerID="8595de5956503317e9c9965ad2288163bbbdf86eb190728548bc8ff6fb456660" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.061015 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.106619 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mk7c\" (UniqueName: \"kubernetes.io/projected/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-kube-api-access-7mk7c\") pod \"nova-cell0-db-create-frhtr\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.106707 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwg92\" (UniqueName: \"kubernetes.io/projected/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-kube-api-access-xwg92\") pod \"nova-api-c90a-account-create-update-24496\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.106793 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l8wx\" (UniqueName: \"kubernetes.io/projected/7ff4c92a-13c5-4a18-944d-38682d75ec07-kube-api-access-7l8wx\") pod \"nova-api-db-create-d4zrc\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.106870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-operator-scripts\") pod \"nova-api-c90a-account-create-update-24496\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.106896 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-operator-scripts\") pod \"nova-cell0-db-create-frhtr\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.106997 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff4c92a-13c5-4a18-944d-38682d75ec07-operator-scripts\") pod \"nova-api-db-create-d4zrc\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.108143 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff4c92a-13c5-4a18-944d-38682d75ec07-operator-scripts\") pod \"nova-api-db-create-d4zrc\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.126415 4758 scope.go:117] "RemoveContainer" containerID="9299b9e27e5198daa91a90f721e702d0ea26aa93e94e1904454efd0443ccf678" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.126726 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c90a-account-create-update-24496"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.134179 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" podStartSLOduration=10.134149513 podStartE2EDuration="10.134149513s" podCreationTimestamp="2025-12-03 17:19:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:04.061615437 +0000 UTC m=+1459.262992298" watchObservedRunningTime="2025-12-03 17:20:04.134149513 +0000 UTC m=+1459.335526374" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.147604 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l8wx\" (UniqueName: \"kubernetes.io/projected/7ff4c92a-13c5-4a18-944d-38682d75ec07-kube-api-access-7l8wx\") pod \"nova-api-db-create-d4zrc\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.198217 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.215308 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwg92\" (UniqueName: \"kubernetes.io/projected/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-kube-api-access-xwg92\") pod \"nova-api-c90a-account-create-update-24496\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.215606 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-operator-scripts\") pod \"nova-api-c90a-account-create-update-24496\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.215673 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-operator-scripts\") pod \"nova-cell0-db-create-frhtr\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.216162 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mk7c\" (UniqueName: \"kubernetes.io/projected/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-kube-api-access-7mk7c\") pod \"nova-cell0-db-create-frhtr\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.218872 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-operator-scripts\") pod \"nova-cell0-db-create-frhtr\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.225774 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-operator-scripts\") pod \"nova-api-c90a-account-create-update-24496\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.248889 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.2926115 podStartE2EDuration="16.248848392s" podCreationTimestamp="2025-12-03 17:19:48 +0000 UTC" firstStartedPulling="2025-12-03 17:19:49.800904056 +0000 UTC m=+1445.002280917" lastFinishedPulling="2025-12-03 17:20:02.757140948 +0000 UTC m=+1457.958517809" observedRunningTime="2025-12-03 17:20:04.097399077 +0000 UTC m=+1459.298775938" watchObservedRunningTime="2025-12-03 17:20:04.248848392 +0000 UTC m=+1459.450225263" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.251953 4758 scope.go:117] "RemoveContainer" containerID="d4fdb1467bc7e1d7119c3cd23b4a6f9c252938f343e72f940f4800aed1ad0cd9" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.266578 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwg92\" (UniqueName: \"kubernetes.io/projected/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-kube-api-access-xwg92\") pod \"nova-api-c90a-account-create-update-24496\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.302703 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.302927 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.313458 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.316180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.317427 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mk7c\" (UniqueName: \"kubernetes.io/projected/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-kube-api-access-7mk7c\") pod \"nova-cell0-db-create-frhtr\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.320319 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.329162 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.329359 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.394801 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xxw46"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.396773 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.418274 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420592 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-config-data\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420770 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-run-httpd\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420885 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-scripts\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420918 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420974 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pllg\" (UniqueName: \"kubernetes.io/projected/10254294-76ba-47ba-9bdb-a4a942dadb18-kube-api-access-9pllg\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.420999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-log-httpd\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.452785 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.496390 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xxw46"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522080 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cncnk\" (UniqueName: \"kubernetes.io/projected/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-kube-api-access-cncnk\") pod \"nova-cell1-db-create-xxw46\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522452 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-run-httpd\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522556 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522700 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-scripts\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522783 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-operator-scripts\") pod \"nova-cell1-db-create-xxw46\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.522971 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pllg\" (UniqueName: \"kubernetes.io/projected/10254294-76ba-47ba-9bdb-a4a942dadb18-kube-api-access-9pllg\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.523082 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-log-httpd\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.523185 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-config-data\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.527863 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-run-httpd\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.532588 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-log-httpd\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.532650 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-700b-account-create-update-8dvwz"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.534180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.535937 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.536603 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-config-data\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.537149 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.541790 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-scripts\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.547289 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.554893 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-700b-account-create-update-8dvwz"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.592946 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pllg\" (UniqueName: \"kubernetes.io/projected/10254294-76ba-47ba-9bdb-a4a942dadb18-kube-api-access-9pllg\") pod \"ceilometer-0\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.648101 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfjjv\" (UniqueName: \"kubernetes.io/projected/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-kube-api-access-vfjjv\") pod \"nova-cell0-700b-account-create-update-8dvwz\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.648792 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cncnk\" (UniqueName: \"kubernetes.io/projected/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-kube-api-access-cncnk\") pod \"nova-cell1-db-create-xxw46\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.649362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-operator-scripts\") pod \"nova-cell1-db-create-xxw46\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.649713 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-operator-scripts\") pod \"nova-cell0-700b-account-create-update-8dvwz\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.674787 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-operator-scripts\") pod \"nova-cell1-db-create-xxw46\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.724988 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cncnk\" (UniqueName: \"kubernetes.io/projected/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-kube-api-access-cncnk\") pod \"nova-cell1-db-create-xxw46\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.782241 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-operator-scripts\") pod \"nova-cell0-700b-account-create-update-8dvwz\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.782571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfjjv\" (UniqueName: \"kubernetes.io/projected/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-kube-api-access-vfjjv\") pod \"nova-cell0-700b-account-create-update-8dvwz\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.791645 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-operator-scripts\") pod \"nova-cell0-700b-account-create-update-8dvwz\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.792698 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.794071 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-a6be-account-create-update-vfdk6"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.804944 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.810545 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.812117 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfjjv\" (UniqueName: \"kubernetes.io/projected/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-kube-api-access-vfjjv\") pod \"nova-cell0-700b-account-create-update-8dvwz\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.821858 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.828113 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a6be-account-create-update-vfdk6"] Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.895411 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.992854 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbr7d\" (UniqueName: \"kubernetes.io/projected/f24e819a-1dee-418a-99cd-885cf9289b48-kube-api-access-tbr7d\") pod \"nova-cell1-a6be-account-create-update-vfdk6\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:04 crc kubenswrapper[4758]: I1203 17:20:04.992921 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f24e819a-1dee-418a-99cd-885cf9289b48-operator-scripts\") pod \"nova-cell1-a6be-account-create-update-vfdk6\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:04.998584 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.027428 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.095530 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbr7d\" (UniqueName: \"kubernetes.io/projected/f24e819a-1dee-418a-99cd-885cf9289b48-kube-api-access-tbr7d\") pod \"nova-cell1-a6be-account-create-update-vfdk6\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.095585 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f24e819a-1dee-418a-99cd-885cf9289b48-operator-scripts\") pod \"nova-cell1-a6be-account-create-update-vfdk6\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.096440 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f24e819a-1dee-418a-99cd-885cf9289b48-operator-scripts\") pod \"nova-cell1-a6be-account-create-update-vfdk6\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.127653 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbr7d\" (UniqueName: \"kubernetes.io/projected/f24e819a-1dee-418a-99cd-885cf9289b48-kube-api-access-tbr7d\") pod \"nova-cell1-a6be-account-create-update-vfdk6\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.166880 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ae7804-bb04-49dd-844c-78bd4de0b1cb" path="/var/lib/kubelet/pods/f3ae7804-bb04-49dd-844c-78bd4de0b1cb/volumes" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.235143 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-frhtr"] Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.295894 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.449007 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-d4zrc"] Dec 03 17:20:05 crc kubenswrapper[4758]: W1203 17:20:05.462775 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8eb3ef8_1b17_4d2d_b664_f32d7263d559.slice/crio-b0a8c04f58bb35a620b72bc09fc066a4095bc3e76ea97c9554aadc60c9736b77 WatchSource:0}: Error finding container b0a8c04f58bb35a620b72bc09fc066a4095bc3e76ea97c9554aadc60c9736b77: Status 404 returned error can't find the container with id b0a8c04f58bb35a620b72bc09fc066a4095bc3e76ea97c9554aadc60c9736b77 Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.488295 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c90a-account-create-update-24496"] Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.877799 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:05 crc kubenswrapper[4758]: W1203 17:20:05.904873 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dea5ef6_5ae7_4aff_a453_dbdcfea7ea6f.slice/crio-6f30337646460692164f7a52e7ac263ff46e06c88ef1c06c89a91f914cb42efd WatchSource:0}: Error finding container 6f30337646460692164f7a52e7ac263ff46e06c88ef1c06c89a91f914cb42efd: Status 404 returned error can't find the container with id 6f30337646460692164f7a52e7ac263ff46e06c88ef1c06c89a91f914cb42efd Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.906451 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": read tcp 10.217.0.2:44138->10.217.0.149:9292: read: connection reset by peer" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.906729 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": read tcp 10.217.0.2:44142->10.217.0.149:9292: read: connection reset by peer" Dec 03 17:20:05 crc kubenswrapper[4758]: I1203 17:20:05.913875 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xxw46"] Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.013320 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-700b-account-create-update-8dvwz"] Dec 03 17:20:06 crc kubenswrapper[4758]: W1203 17:20:06.041552 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97d849bd_1f04_4a59_805a_cb9ae0f3b7bd.slice/crio-a0427920706450753f4a747ff72d3864998e8d85784e4d8687f8055c4e49dc2a WatchSource:0}: Error finding container a0427920706450753f4a747ff72d3864998e8d85784e4d8687f8055c4e49dc2a: Status 404 returned error can't find the container with id a0427920706450753f4a747ff72d3864998e8d85784e4d8687f8055c4e49dc2a Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.091902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c90a-account-create-update-24496" event={"ID":"b8eb3ef8-1b17-4d2d-b664-f32d7263d559","Type":"ContainerStarted","Data":"b0a8c04f58bb35a620b72bc09fc066a4095bc3e76ea97c9554aadc60c9736b77"} Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.096076 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxw46" event={"ID":"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f","Type":"ContainerStarted","Data":"6f30337646460692164f7a52e7ac263ff46e06c88ef1c06c89a91f914cb42efd"} Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.099511 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d4zrc" event={"ID":"7ff4c92a-13c5-4a18-944d-38682d75ec07","Type":"ContainerStarted","Data":"7e4e94febc3552f3886fb0b5c4dd24e392497e90a5ebc59b5ca3974f177fed36"} Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.104646 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" event={"ID":"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd","Type":"ContainerStarted","Data":"a0427920706450753f4a747ff72d3864998e8d85784e4d8687f8055c4e49dc2a"} Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.105927 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frhtr" event={"ID":"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c","Type":"ContainerStarted","Data":"d5941e98608dc31b75b5d4250f5fe4dd55b264056af0483dfd6d308d03984751"} Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.119435 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerStarted","Data":"4759e581bae15fe3b6931e7a6a45784f38e2bddc9c95cb0df240e8a07b73d1d5"} Dec 03 17:20:06 crc kubenswrapper[4758]: I1203 17:20:06.271027 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a6be-account-create-update-vfdk6"] Dec 03 17:20:07 crc kubenswrapper[4758]: I1203 17:20:07.134719 4758 generic.go:334] "Generic (PLEG): container finished" podID="74ca64c8-28c7-472e-85bd-acd6af307580" containerID="ea6dcf9c2fc3a2c5e5cd9cbbf22a997692e23bc0e78284fb1d6304f44a979e57" exitCode=0 Dec 03 17:20:07 crc kubenswrapper[4758]: I1203 17:20:07.134789 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ca64c8-28c7-472e-85bd-acd6af307580","Type":"ContainerDied","Data":"ea6dcf9c2fc3a2c5e5cd9cbbf22a997692e23bc0e78284fb1d6304f44a979e57"} Dec 03 17:20:07 crc kubenswrapper[4758]: I1203 17:20:07.864371 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:20:07 crc kubenswrapper[4758]: I1203 17:20:07.865220 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-log" containerID="cri-o://cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d" gracePeriod=30 Dec 03 17:20:07 crc kubenswrapper[4758]: I1203 17:20:07.865794 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-httpd" containerID="cri-o://93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d" gracePeriod=30 Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.156984 4758 generic.go:334] "Generic (PLEG): container finished" podID="12d6c948-8253-4dbf-b782-27603b7ae061" containerID="cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d" exitCode=143 Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.157372 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"12d6c948-8253-4dbf-b782-27603b7ae061","Type":"ContainerDied","Data":"cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.159354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerStarted","Data":"2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.160358 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" event={"ID":"f24e819a-1dee-418a-99cd-885cf9289b48","Type":"ContainerStarted","Data":"d747aff75568811a8371fabd136f0f4d2fe3a3d7d5d4dd1c649ff970432797c5"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.160380 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" event={"ID":"f24e819a-1dee-418a-99cd-885cf9289b48","Type":"ContainerStarted","Data":"ff8033c0aa1273b27778f7f6ff7a9ba76a47058ea8d3c56f5e8cd0c5eb565bd5"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.163299 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c90a-account-create-update-24496" event={"ID":"b8eb3ef8-1b17-4d2d-b664-f32d7263d559","Type":"ContainerStarted","Data":"f2582536bf237b4937da05dde1ec781a6082f8afc2d3f323fba8ffc0d87b12ca"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.166809 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxw46" event={"ID":"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f","Type":"ContainerStarted","Data":"c9a053a98293b8d5dd7c40e6e5514af1ae1e974be1223ffac53d82fad7efd74a"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.169839 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d4zrc" event={"ID":"7ff4c92a-13c5-4a18-944d-38682d75ec07","Type":"ContainerStarted","Data":"b9a6a87f277fcfa9d46702f8fb08b9dd826f5798578c0b34f74e4508f4c77b21"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.172396 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" event={"ID":"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd","Type":"ContainerStarted","Data":"e5480ee329143645fdfee867011988a34e8d68ff7a64b40999efdb3782b3f44d"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.177561 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frhtr" event={"ID":"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c","Type":"ContainerStarted","Data":"7277037e4a4d38c2ea41fe5cf0960dceaeccfafcabb4c1f19da9780663d3a688"} Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.189230 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" podStartSLOduration=4.189202243 podStartE2EDuration="4.189202243s" podCreationTimestamp="2025-12-03 17:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:08.181509277 +0000 UTC m=+1463.382886138" watchObservedRunningTime="2025-12-03 17:20:08.189202243 +0000 UTC m=+1463.390579094" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.199358 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c90a-account-create-update-24496" podStartSLOduration=5.199339496 podStartE2EDuration="5.199339496s" podCreationTimestamp="2025-12-03 17:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:08.198459542 +0000 UTC m=+1463.399836403" watchObservedRunningTime="2025-12-03 17:20:08.199339496 +0000 UTC m=+1463.400716357" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.222653 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-xxw46" podStartSLOduration=4.22262689 podStartE2EDuration="4.22262689s" podCreationTimestamp="2025-12-03 17:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:08.222586449 +0000 UTC m=+1463.423963300" watchObservedRunningTime="2025-12-03 17:20:08.22262689 +0000 UTC m=+1463.424003751" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.267283 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-frhtr" podStartSLOduration=5.267249868 podStartE2EDuration="5.267249868s" podCreationTimestamp="2025-12-03 17:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:08.252022719 +0000 UTC m=+1463.453399580" watchObservedRunningTime="2025-12-03 17:20:08.267249868 +0000 UTC m=+1463.468626729" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.279266 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-d4zrc" podStartSLOduration=5.27924497 podStartE2EDuration="5.27924497s" podCreationTimestamp="2025-12-03 17:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:08.277200145 +0000 UTC m=+1463.478577016" watchObservedRunningTime="2025-12-03 17:20:08.27924497 +0000 UTC m=+1463.480621831" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.303122 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" podStartSLOduration=4.30308671 podStartE2EDuration="4.30308671s" podCreationTimestamp="2025-12-03 17:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:08.297590563 +0000 UTC m=+1463.498967424" watchObservedRunningTime="2025-12-03 17:20:08.30308671 +0000 UTC m=+1463.504463571" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.436860 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.844651 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:20:08 crc kubenswrapper[4758]: E1203 17:20:08.886000 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf24e819a_1dee_418a_99cd_885cf9289b48.slice/crio-conmon-d747aff75568811a8371fabd136f0f4d2fe3a3d7d5d4dd1c649ff970432797c5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc2e397d_cdcc_4116_8dfb_5b3ac6da924c.slice/crio-7277037e4a4d38c2ea41fe5cf0960dceaeccfafcabb4c1f19da9780663d3a688.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957135 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8bf5\" (UniqueName: \"kubernetes.io/projected/74ca64c8-28c7-472e-85bd-acd6af307580-kube-api-access-n8bf5\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957197 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-logs\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-scripts\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957297 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957339 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-config-data\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957406 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-public-tls-certs\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957430 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-httpd-run\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.957480 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-combined-ca-bundle\") pod \"74ca64c8-28c7-472e-85bd-acd6af307580\" (UID: \"74ca64c8-28c7-472e-85bd-acd6af307580\") " Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.961901 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.962185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-logs" (OuterVolumeSpecName: "logs") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.973429 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74ca64c8-28c7-472e-85bd-acd6af307580-kube-api-access-n8bf5" (OuterVolumeSpecName: "kube-api-access-n8bf5") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "kube-api-access-n8bf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.974606 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-scripts" (OuterVolumeSpecName: "scripts") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4758]: I1203 17:20:08.975139 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.031943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065669 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065881 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8bf5\" (UniqueName: \"kubernetes.io/projected/74ca64c8-28c7-472e-85bd-acd6af307580-kube-api-access-n8bf5\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065904 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065917 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065949 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065962 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065973 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ca64c8-28c7-472e-85bd-acd6af307580-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.065983 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.092544 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-config-data" (OuterVolumeSpecName: "config-data") pod "74ca64c8-28c7-472e-85bd-acd6af307580" (UID: "74ca64c8-28c7-472e-85bd-acd6af307580"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.108596 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.167736 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.167796 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ca64c8-28c7-472e-85bd-acd6af307580-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.190704 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerStarted","Data":"74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.193307 4758 generic.go:334] "Generic (PLEG): container finished" podID="f24e819a-1dee-418a-99cd-885cf9289b48" containerID="d747aff75568811a8371fabd136f0f4d2fe3a3d7d5d4dd1c649ff970432797c5" exitCode=0 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.193365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" event={"ID":"f24e819a-1dee-418a-99cd-885cf9289b48","Type":"ContainerDied","Data":"d747aff75568811a8371fabd136f0f4d2fe3a3d7d5d4dd1c649ff970432797c5"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.195769 4758 generic.go:334] "Generic (PLEG): container finished" podID="b8eb3ef8-1b17-4d2d-b664-f32d7263d559" containerID="f2582536bf237b4937da05dde1ec781a6082f8afc2d3f323fba8ffc0d87b12ca" exitCode=0 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.195823 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c90a-account-create-update-24496" event={"ID":"b8eb3ef8-1b17-4d2d-b664-f32d7263d559","Type":"ContainerDied","Data":"f2582536bf237b4937da05dde1ec781a6082f8afc2d3f323fba8ffc0d87b12ca"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.197475 4758 generic.go:334] "Generic (PLEG): container finished" podID="6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" containerID="c9a053a98293b8d5dd7c40e6e5514af1ae1e974be1223ffac53d82fad7efd74a" exitCode=0 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.197519 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxw46" event={"ID":"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f","Type":"ContainerDied","Data":"c9a053a98293b8d5dd7c40e6e5514af1ae1e974be1223ffac53d82fad7efd74a"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.199771 4758 generic.go:334] "Generic (PLEG): container finished" podID="7ff4c92a-13c5-4a18-944d-38682d75ec07" containerID="b9a6a87f277fcfa9d46702f8fb08b9dd826f5798578c0b34f74e4508f4c77b21" exitCode=0 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.199856 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d4zrc" event={"ID":"7ff4c92a-13c5-4a18-944d-38682d75ec07","Type":"ContainerDied","Data":"b9a6a87f277fcfa9d46702f8fb08b9dd826f5798578c0b34f74e4508f4c77b21"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.201289 4758 generic.go:334] "Generic (PLEG): container finished" podID="97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" containerID="e5480ee329143645fdfee867011988a34e8d68ff7a64b40999efdb3782b3f44d" exitCode=0 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.201349 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" event={"ID":"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd","Type":"ContainerDied","Data":"e5480ee329143645fdfee867011988a34e8d68ff7a64b40999efdb3782b3f44d"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.203105 4758 generic.go:334] "Generic (PLEG): container finished" podID="dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" containerID="7277037e4a4d38c2ea41fe5cf0960dceaeccfafcabb4c1f19da9780663d3a688" exitCode=0 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.203168 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frhtr" event={"ID":"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c","Type":"ContainerDied","Data":"7277037e4a4d38c2ea41fe5cf0960dceaeccfafcabb4c1f19da9780663d3a688"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.206481 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ca64c8-28c7-472e-85bd-acd6af307580","Type":"ContainerDied","Data":"89b58a21b26fe2a2527553d5556f7d694876e49729ca2f5d4fe36850040209d1"} Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.206543 4758 scope.go:117] "RemoveContainer" containerID="ea6dcf9c2fc3a2c5e5cd9cbbf22a997692e23bc0e78284fb1d6304f44a979e57" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.206597 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.273911 4758 scope.go:117] "RemoveContainer" containerID="55c59d419d0dddb9e20a6a9054a34c3b529a7c39deb72363a316ff284231e1fe" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.375834 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.412196 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.420050 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:20:09 crc kubenswrapper[4758]: E1203 17:20:09.420666 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-log" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.420803 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-log" Dec 03 17:20:09 crc kubenswrapper[4758]: E1203 17:20:09.420857 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-httpd" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.420865 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-httpd" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.421134 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-httpd" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.421169 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" containerName="glance-log" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.422547 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.429537 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.429957 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.460974 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581147 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581255 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-logs\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581282 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581311 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581436 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-scripts\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581544 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9gs4\" (UniqueName: \"kubernetes.io/projected/be629420-0e78-44d4-9ca1-4ee1601d4ca6-kube-api-access-b9gs4\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581730 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.581816 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-config-data\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.618822 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.684813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.684992 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-config-data\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.685184 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.685274 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-logs\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.685365 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.685444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.685571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-scripts\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.685738 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9gs4\" (UniqueName: \"kubernetes.io/projected/be629420-0e78-44d4-9ca1-4ee1601d4ca6-kube-api-access-b9gs4\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.686518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-logs\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.686930 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.700194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.710828 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.719327 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-scripts\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.725982 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9gs4\" (UniqueName: \"kubernetes.io/projected/be629420-0e78-44d4-9ca1-4ee1601d4ca6-kube-api-access-b9gs4\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.728462 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-config-data\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.730655 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.746516 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64f4d5ff96-vfvv7"] Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.747191 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64f4d5ff96-vfvv7" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-api" containerID="cri-o://0b6b8750178fa432f12fa07d0ae26570d538a6021da9fde9be6813c446d4d7ee" gracePeriod=30 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.747418 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64f4d5ff96-vfvv7" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-httpd" containerID="cri-o://34795b0ebc5422c2fc4ff2442f6fa1ac811ee787cb0ea6513b2b2de3fde194d6" gracePeriod=30 Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.752111 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " pod="openstack/glance-default-external-api-0" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.983513 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:20:09 crc kubenswrapper[4758]: I1203 17:20:09.985412 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:20:10 crc kubenswrapper[4758]: I1203 17:20:10.052408 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:20:10 crc kubenswrapper[4758]: I1203 17:20:10.237825 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerStarted","Data":"aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f"} Dec 03 17:20:10 crc kubenswrapper[4758]: I1203 17:20:10.244448 4758 generic.go:334] "Generic (PLEG): container finished" podID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerID="34795b0ebc5422c2fc4ff2442f6fa1ac811ee787cb0ea6513b2b2de3fde194d6" exitCode=0 Dec 03 17:20:10 crc kubenswrapper[4758]: I1203 17:20:10.244836 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f4d5ff96-vfvv7" event={"ID":"d0887aea-e60c-4517-9c8c-c96498f912a4","Type":"ContainerDied","Data":"34795b0ebc5422c2fc4ff2442f6fa1ac811ee787cb0ea6513b2b2de3fde194d6"} Dec 03 17:20:10 crc kubenswrapper[4758]: I1203 17:20:10.848300 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:10 crc kubenswrapper[4758]: I1203 17:20:10.863782 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.050671 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwg92\" (UniqueName: \"kubernetes.io/projected/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-kube-api-access-xwg92\") pod \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.051278 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l8wx\" (UniqueName: \"kubernetes.io/projected/7ff4c92a-13c5-4a18-944d-38682d75ec07-kube-api-access-7l8wx\") pod \"7ff4c92a-13c5-4a18-944d-38682d75ec07\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.051318 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-operator-scripts\") pod \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\" (UID: \"b8eb3ef8-1b17-4d2d-b664-f32d7263d559\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.051390 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff4c92a-13c5-4a18-944d-38682d75ec07-operator-scripts\") pod \"7ff4c92a-13c5-4a18-944d-38682d75ec07\" (UID: \"7ff4c92a-13c5-4a18-944d-38682d75ec07\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.052881 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8eb3ef8-1b17-4d2d-b664-f32d7263d559" (UID: "b8eb3ef8-1b17-4d2d-b664-f32d7263d559"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.058589 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-kube-api-access-xwg92" (OuterVolumeSpecName: "kube-api-access-xwg92") pod "b8eb3ef8-1b17-4d2d-b664-f32d7263d559" (UID: "b8eb3ef8-1b17-4d2d-b664-f32d7263d559"). InnerVolumeSpecName "kube-api-access-xwg92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.069038 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff4c92a-13c5-4a18-944d-38682d75ec07-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ff4c92a-13c5-4a18-944d-38682d75ec07" (UID: "7ff4c92a-13c5-4a18-944d-38682d75ec07"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.072027 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff4c92a-13c5-4a18-944d-38682d75ec07-kube-api-access-7l8wx" (OuterVolumeSpecName: "kube-api-access-7l8wx") pod "7ff4c92a-13c5-4a18-944d-38682d75ec07" (UID: "7ff4c92a-13c5-4a18-944d-38682d75ec07"). InnerVolumeSpecName "kube-api-access-7l8wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.158895 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwg92\" (UniqueName: \"kubernetes.io/projected/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-kube-api-access-xwg92\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.158932 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l8wx\" (UniqueName: \"kubernetes.io/projected/7ff4c92a-13c5-4a18-944d-38682d75ec07-kube-api-access-7l8wx\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.158941 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8eb3ef8-1b17-4d2d-b664-f32d7263d559-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.158951 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff4c92a-13c5-4a18-944d-38682d75ec07-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.166165 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74ca64c8-28c7-472e-85bd-acd6af307580" path="/var/lib/kubelet/pods/74ca64c8-28c7-472e-85bd-acd6af307580/volumes" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.208317 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.294804 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" event={"ID":"f24e819a-1dee-418a-99cd-885cf9289b48","Type":"ContainerDied","Data":"ff8033c0aa1273b27778f7f6ff7a9ba76a47058ea8d3c56f5e8cd0c5eb565bd5"} Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.294872 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff8033c0aa1273b27778f7f6ff7a9ba76a47058ea8d3c56f5e8cd0c5eb565bd5" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.294974 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a6be-account-create-update-vfdk6" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.305719 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c90a-account-create-update-24496" event={"ID":"b8eb3ef8-1b17-4d2d-b664-f32d7263d559","Type":"ContainerDied","Data":"b0a8c04f58bb35a620b72bc09fc066a4095bc3e76ea97c9554aadc60c9736b77"} Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.305775 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0a8c04f58bb35a620b72bc09fc066a4095bc3e76ea97c9554aadc60c9736b77" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.305865 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c90a-account-create-update-24496" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.319558 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxw46" event={"ID":"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f","Type":"ContainerDied","Data":"6f30337646460692164f7a52e7ac263ff46e06c88ef1c06c89a91f914cb42efd"} Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.319646 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f30337646460692164f7a52e7ac263ff46e06c88ef1c06c89a91f914cb42efd" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.325995 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d4zrc" event={"ID":"7ff4c92a-13c5-4a18-944d-38682d75ec07","Type":"ContainerDied","Data":"7e4e94febc3552f3886fb0b5c4dd24e392497e90a5ebc59b5ca3974f177fed36"} Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.326071 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e4e94febc3552f3886fb0b5c4dd24e392497e90a5ebc59b5ca3974f177fed36" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.326178 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d4zrc" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.329571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frhtr" event={"ID":"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c","Type":"ContainerDied","Data":"d5941e98608dc31b75b5d4250f5fe4dd55b264056af0483dfd6d308d03984751"} Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.329845 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5941e98608dc31b75b5d4250f5fe4dd55b264056af0483dfd6d308d03984751" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.363805 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f24e819a-1dee-418a-99cd-885cf9289b48-operator-scripts\") pod \"f24e819a-1dee-418a-99cd-885cf9289b48\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.364083 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbr7d\" (UniqueName: \"kubernetes.io/projected/f24e819a-1dee-418a-99cd-885cf9289b48-kube-api-access-tbr7d\") pod \"f24e819a-1dee-418a-99cd-885cf9289b48\" (UID: \"f24e819a-1dee-418a-99cd-885cf9289b48\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.366107 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f24e819a-1dee-418a-99cd-885cf9289b48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f24e819a-1dee-418a-99cd-885cf9289b48" (UID: "f24e819a-1dee-418a-99cd-885cf9289b48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.370774 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f24e819a-1dee-418a-99cd-885cf9289b48-kube-api-access-tbr7d" (OuterVolumeSpecName: "kube-api-access-tbr7d") pod "f24e819a-1dee-418a-99cd-885cf9289b48" (UID: "f24e819a-1dee-418a-99cd-885cf9289b48"). InnerVolumeSpecName "kube-api-access-tbr7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.466753 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f24e819a-1dee-418a-99cd-885cf9289b48-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.467397 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbr7d\" (UniqueName: \"kubernetes.io/projected/f24e819a-1dee-418a-99cd-885cf9289b48-kube-api-access-tbr7d\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.509607 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.526433 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.623773 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.630625 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.671714 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cncnk\" (UniqueName: \"kubernetes.io/projected/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-kube-api-access-cncnk\") pod \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.671865 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-operator-scripts\") pod \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\" (UID: \"6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.673632 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" (UID: "6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.689511 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-kube-api-access-cncnk" (OuterVolumeSpecName: "kube-api-access-cncnk") pod "6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" (UID: "6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f"). InnerVolumeSpecName "kube-api-access-cncnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.773595 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-operator-scripts\") pod \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.773763 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-operator-scripts\") pod \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.773822 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfjjv\" (UniqueName: \"kubernetes.io/projected/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-kube-api-access-vfjjv\") pod \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\" (UID: \"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.773895 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mk7c\" (UniqueName: \"kubernetes.io/projected/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-kube-api-access-7mk7c\") pod \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\" (UID: \"dc2e397d-cdcc-4116-8dfb-5b3ac6da924c\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.774448 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cncnk\" (UniqueName: \"kubernetes.io/projected/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-kube-api-access-cncnk\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.774460 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.774626 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" (UID: "dc2e397d-cdcc-4116-8dfb-5b3ac6da924c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.775099 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" (UID: "97d849bd-1f04-4a59-805a-cb9ae0f3b7bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.778900 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-kube-api-access-7mk7c" (OuterVolumeSpecName: "kube-api-access-7mk7c") pod "dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" (UID: "dc2e397d-cdcc-4116-8dfb-5b3ac6da924c"). InnerVolumeSpecName "kube-api-access-7mk7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.784402 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-kube-api-access-vfjjv" (OuterVolumeSpecName: "kube-api-access-vfjjv") pod "97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" (UID: "97d849bd-1f04-4a59-805a-cb9ae0f3b7bd"). InnerVolumeSpecName "kube-api-access-vfjjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.877708 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.877752 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.877763 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfjjv\" (UniqueName: \"kubernetes.io/projected/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd-kube-api-access-vfjjv\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.877776 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mk7c\" (UniqueName: \"kubernetes.io/projected/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c-kube-api-access-7mk7c\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.940656 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.984898 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-httpd-run\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.984958 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-scripts\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.985034 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-internal-tls-certs\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.985092 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tllxp\" (UniqueName: \"kubernetes.io/projected/12d6c948-8253-4dbf-b782-27603b7ae061-kube-api-access-tllxp\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.985182 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-logs\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.985275 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-config-data\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.985306 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-combined-ca-bundle\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.985838 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"12d6c948-8253-4dbf-b782-27603b7ae061\" (UID: \"12d6c948-8253-4dbf-b782-27603b7ae061\") " Dec 03 17:20:11 crc kubenswrapper[4758]: I1203 17:20:11.997693 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.000147 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.000519 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-logs" (OuterVolumeSpecName: "logs") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.010748 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-scripts" (OuterVolumeSpecName: "scripts") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.010832 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d6c948-8253-4dbf-b782-27603b7ae061-kube-api-access-tllxp" (OuterVolumeSpecName: "kube-api-access-tllxp") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "kube-api-access-tllxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.036239 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.083536 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-config-data" (OuterVolumeSpecName: "config-data") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089375 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tllxp\" (UniqueName: \"kubernetes.io/projected/12d6c948-8253-4dbf-b782-27603b7ae061-kube-api-access-tllxp\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089427 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089443 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089458 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089489 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089502 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d6c948-8253-4dbf-b782-27603b7ae061-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.089513 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.125288 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.148577 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "12d6c948-8253-4dbf-b782-27603b7ae061" (UID: "12d6c948-8253-4dbf-b782-27603b7ae061"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.205743 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.205897 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d6c948-8253-4dbf-b782-27603b7ae061-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.373109 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" event={"ID":"97d849bd-1f04-4a59-805a-cb9ae0f3b7bd","Type":"ContainerDied","Data":"a0427920706450753f4a747ff72d3864998e8d85784e4d8687f8055c4e49dc2a"} Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.373172 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0427920706450753f4a747ff72d3864998e8d85784e4d8687f8055c4e49dc2a" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.373274 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-700b-account-create-update-8dvwz" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.392223 4758 generic.go:334] "Generic (PLEG): container finished" podID="12d6c948-8253-4dbf-b782-27603b7ae061" containerID="93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d" exitCode=0 Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.392343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"12d6c948-8253-4dbf-b782-27603b7ae061","Type":"ContainerDied","Data":"93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d"} Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.392382 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"12d6c948-8253-4dbf-b782-27603b7ae061","Type":"ContainerDied","Data":"e6d6992da89cfbc258334346183f4d711c90d32212a8532b19cbfb896282e30e"} Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.392405 4758 scope.go:117] "RemoveContainer" containerID="93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.392614 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.402328 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be629420-0e78-44d4-9ca1-4ee1601d4ca6","Type":"ContainerStarted","Data":"a416b337d477e7ddcf7f04a362cfc7fd6501517331e259aa030296f63e2f24c5"} Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409206 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frhtr" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409263 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerStarted","Data":"295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35"} Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409321 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxw46" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409313 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-central-agent" containerID="cri-o://2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5" gracePeriod=30 Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409466 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="proxy-httpd" containerID="cri-o://295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35" gracePeriod=30 Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409508 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409527 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-notification-agent" containerID="cri-o://74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8" gracePeriod=30 Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.409512 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="sg-core" containerID="cri-o://aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f" gracePeriod=30 Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.441093 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.34844764 podStartE2EDuration="8.441058396s" podCreationTimestamp="2025-12-03 17:20:04 +0000 UTC" firstStartedPulling="2025-12-03 17:20:05.905250536 +0000 UTC m=+1461.106627397" lastFinishedPulling="2025-12-03 17:20:10.997861292 +0000 UTC m=+1466.199238153" observedRunningTime="2025-12-03 17:20:12.437151721 +0000 UTC m=+1467.638528592" watchObservedRunningTime="2025-12-03 17:20:12.441058396 +0000 UTC m=+1467.642435257" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.466995 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.489193 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506028 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506718 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8eb3ef8-1b17-4d2d-b664-f32d7263d559" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506743 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8eb3ef8-1b17-4d2d-b664-f32d7263d559" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506768 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506778 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506801 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506809 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506821 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-log" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506829 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-log" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506842 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f24e819a-1dee-418a-99cd-885cf9289b48" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506850 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f24e819a-1dee-418a-99cd-885cf9289b48" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506862 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-httpd" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506869 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-httpd" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506900 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff4c92a-13c5-4a18-944d-38682d75ec07" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506908 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff4c92a-13c5-4a18-944d-38682d75ec07" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.506925 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.506934 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507152 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507166 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-log" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507180 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507198 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" containerName="glance-httpd" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507210 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff4c92a-13c5-4a18-944d-38682d75ec07" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507228 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" containerName="mariadb-database-create" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507242 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f24e819a-1dee-418a-99cd-885cf9289b48" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.507257 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8eb3ef8-1b17-4d2d-b664-f32d7263d559" containerName="mariadb-account-create-update" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.508937 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.512304 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.514342 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.550098 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.615791 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg4gr\" (UniqueName: \"kubernetes.io/projected/1b49b1d3-cbb4-49da-8244-69e15595da33-kube-api-access-mg4gr\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.615864 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.615914 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.615983 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.616010 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.616041 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.616069 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.616105 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.718581 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg4gr\" (UniqueName: \"kubernetes.io/projected/1b49b1d3-cbb4-49da-8244-69e15595da33-kube-api-access-mg4gr\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.718673 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.718907 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.718900 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.720551 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.720623 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.720699 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.720754 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.720836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.721762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.722501 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.733465 4758 scope.go:117] "RemoveContainer" containerID="cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.734421 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.739225 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.743315 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.755827 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.761902 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg4gr\" (UniqueName: \"kubernetes.io/projected/1b49b1d3-cbb4-49da-8244-69e15595da33-kube-api-access-mg4gr\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.788319 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " pod="openstack/glance-default-internal-api-0" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.870879 4758 scope.go:117] "RemoveContainer" containerID="93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.871477 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d\": container with ID starting with 93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d not found: ID does not exist" containerID="93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.871531 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d"} err="failed to get container status \"93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d\": rpc error: code = NotFound desc = could not find container \"93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d\": container with ID starting with 93e03265e5cb6a4107562cf9a0861cf1d0de1112bc540a685a644a077fd8017d not found: ID does not exist" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.871793 4758 scope.go:117] "RemoveContainer" containerID="cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d" Dec 03 17:20:12 crc kubenswrapper[4758]: E1203 17:20:12.872211 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d\": container with ID starting with cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d not found: ID does not exist" containerID="cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.872283 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d"} err="failed to get container status \"cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d\": rpc error: code = NotFound desc = could not find container \"cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d\": container with ID starting with cd1daba29e6f6829cb8cd62ca3254188215cce9d2c063d951a4ddbebc830557d not found: ID does not exist" Dec 03 17:20:12 crc kubenswrapper[4758]: I1203 17:20:12.878545 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.158409 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d6c948-8253-4dbf-b782-27603b7ae061" path="/var/lib/kubelet/pods/12d6c948-8253-4dbf-b782-27603b7ae061/volumes" Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.439042 4758 generic.go:334] "Generic (PLEG): container finished" podID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerID="295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35" exitCode=0 Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.439102 4758 generic.go:334] "Generic (PLEG): container finished" podID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerID="aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f" exitCode=2 Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.439112 4758 generic.go:334] "Generic (PLEG): container finished" podID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerID="74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8" exitCode=0 Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.439139 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerDied","Data":"295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35"} Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.439231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerDied","Data":"aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f"} Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.439251 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerDied","Data":"74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8"} Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.443854 4758 generic.go:334] "Generic (PLEG): container finished" podID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerID="0b6b8750178fa432f12fa07d0ae26570d538a6021da9fde9be6813c446d4d7ee" exitCode=0 Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.443953 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f4d5ff96-vfvv7" event={"ID":"d0887aea-e60c-4517-9c8c-c96498f912a4","Type":"ContainerDied","Data":"0b6b8750178fa432f12fa07d0ae26570d538a6021da9fde9be6813c446d4d7ee"} Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.452534 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be629420-0e78-44d4-9ca1-4ee1601d4ca6","Type":"ContainerStarted","Data":"b86b8be6d1e76610577e86e34e754975aa88fa3ee8f7d13944550ef42f337ec8"} Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.616079 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:20:13 crc kubenswrapper[4758]: W1203 17:20:13.635376 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b49b1d3_cbb4_49da_8244_69e15595da33.slice/crio-42229c66216558885e65ed8defa507e5e594d94a41d371a6cf2e4e46a13a6971 WatchSource:0}: Error finding container 42229c66216558885e65ed8defa507e5e594d94a41d371a6cf2e4e46a13a6971: Status 404 returned error can't find the container with id 42229c66216558885e65ed8defa507e5e594d94a41d371a6cf2e4e46a13a6971 Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.800292 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.975834 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-ovndb-tls-certs\") pod \"d0887aea-e60c-4517-9c8c-c96498f912a4\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.975913 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-config\") pod \"d0887aea-e60c-4517-9c8c-c96498f912a4\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.976071 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-httpd-config\") pod \"d0887aea-e60c-4517-9c8c-c96498f912a4\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.976212 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvcsb\" (UniqueName: \"kubernetes.io/projected/d0887aea-e60c-4517-9c8c-c96498f912a4-kube-api-access-dvcsb\") pod \"d0887aea-e60c-4517-9c8c-c96498f912a4\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.976238 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-combined-ca-bundle\") pod \"d0887aea-e60c-4517-9c8c-c96498f912a4\" (UID: \"d0887aea-e60c-4517-9c8c-c96498f912a4\") " Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.982845 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d0887aea-e60c-4517-9c8c-c96498f912a4" (UID: "d0887aea-e60c-4517-9c8c-c96498f912a4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:13 crc kubenswrapper[4758]: I1203 17:20:13.989395 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0887aea-e60c-4517-9c8c-c96498f912a4-kube-api-access-dvcsb" (OuterVolumeSpecName: "kube-api-access-dvcsb") pod "d0887aea-e60c-4517-9c8c-c96498f912a4" (UID: "d0887aea-e60c-4517-9c8c-c96498f912a4"). InnerVolumeSpecName "kube-api-access-dvcsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.071262 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0887aea-e60c-4517-9c8c-c96498f912a4" (UID: "d0887aea-e60c-4517-9c8c-c96498f912a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.079324 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.079366 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvcsb\" (UniqueName: \"kubernetes.io/projected/d0887aea-e60c-4517-9c8c-c96498f912a4-kube-api-access-dvcsb\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.079380 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.092709 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-config" (OuterVolumeSpecName: "config") pod "d0887aea-e60c-4517-9c8c-c96498f912a4" (UID: "d0887aea-e60c-4517-9c8c-c96498f912a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.127914 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d0887aea-e60c-4517-9c8c-c96498f912a4" (UID: "d0887aea-e60c-4517-9c8c-c96498f912a4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.181528 4758 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.181568 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0887aea-e60c-4517-9c8c-c96498f912a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.469233 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f4d5ff96-vfvv7" event={"ID":"d0887aea-e60c-4517-9c8c-c96498f912a4","Type":"ContainerDied","Data":"8681e2c61ea1c18864ee8448a2f74d8ac7cc45bef5401d9e5ac9eaa8c1babd5d"} Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.469771 4758 scope.go:117] "RemoveContainer" containerID="34795b0ebc5422c2fc4ff2442f6fa1ac811ee787cb0ea6513b2b2de3fde194d6" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.470089 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f4d5ff96-vfvv7" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.478960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be629420-0e78-44d4-9ca1-4ee1601d4ca6","Type":"ContainerStarted","Data":"9d2c93db5d7424078ddb1bb905ea651dea2607f1f94df3e14e8eaf9f6eec920f"} Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.483666 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b49b1d3-cbb4-49da-8244-69e15595da33","Type":"ContainerStarted","Data":"aacdb99dee5e45686e33def0624769051138394c943550e16068c2601899b7a4"} Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.483842 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b49b1d3-cbb4-49da-8244-69e15595da33","Type":"ContainerStarted","Data":"42229c66216558885e65ed8defa507e5e594d94a41d371a6cf2e4e46a13a6971"} Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.516092 4758 scope.go:117] "RemoveContainer" containerID="0b6b8750178fa432f12fa07d0ae26570d538a6021da9fde9be6813c446d4d7ee" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.519983 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.5199616 podStartE2EDuration="5.5199616s" podCreationTimestamp="2025-12-03 17:20:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:14.504252368 +0000 UTC m=+1469.705629229" watchObservedRunningTime="2025-12-03 17:20:14.5199616 +0000 UTC m=+1469.721338481" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.551834 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64f4d5ff96-vfvv7"] Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.563071 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-64f4d5ff96-vfvv7"] Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.912389 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9tms7"] Dec 03 17:20:14 crc kubenswrapper[4758]: E1203 17:20:14.913666 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-httpd" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.913707 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-httpd" Dec 03 17:20:14 crc kubenswrapper[4758]: E1203 17:20:14.913724 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-api" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.913734 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-api" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.914007 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-httpd" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.914027 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" containerName="neutron-api" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.914980 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.917733 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.919313 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6fv2h" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.919461 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 17:20:14 crc kubenswrapper[4758]: I1203 17:20:14.959351 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9tms7"] Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.037017 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.037071 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-scripts\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.037131 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk7cj\" (UniqueName: \"kubernetes.io/projected/52cd754b-1015-487b-a51c-4d7117cfae33-kube-api-access-pk7cj\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.037164 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-config-data\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.065967 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.141060 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0887aea-e60c-4517-9c8c-c96498f912a4" path="/var/lib/kubelet/pods/d0887aea-e60c-4517-9c8c-c96498f912a4/volumes" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.141165 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pllg\" (UniqueName: \"kubernetes.io/projected/10254294-76ba-47ba-9bdb-a4a942dadb18-kube-api-access-9pllg\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.141384 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-log-httpd\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.141440 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-config-data\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.141494 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-run-httpd\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.142016 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.142052 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-scripts\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.142182 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk7cj\" (UniqueName: \"kubernetes.io/projected/52cd754b-1015-487b-a51c-4d7117cfae33-kube-api-access-pk7cj\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.142222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-config-data\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.145028 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.145198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.153517 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.153859 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-config-data\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.161035 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-scripts\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.166207 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10254294-76ba-47ba-9bdb-a4a942dadb18-kube-api-access-9pllg" (OuterVolumeSpecName: "kube-api-access-9pllg") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "kube-api-access-9pllg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.169781 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk7cj\" (UniqueName: \"kubernetes.io/projected/52cd754b-1015-487b-a51c-4d7117cfae33-kube-api-access-pk7cj\") pod \"nova-cell0-conductor-db-sync-9tms7\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.244824 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-scripts\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.244921 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-sg-core-conf-yaml\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.244967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-combined-ca-bundle\") pod \"10254294-76ba-47ba-9bdb-a4a942dadb18\" (UID: \"10254294-76ba-47ba-9bdb-a4a942dadb18\") " Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.247614 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pllg\" (UniqueName: \"kubernetes.io/projected/10254294-76ba-47ba-9bdb-a4a942dadb18-kube-api-access-9pllg\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.248002 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.248016 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10254294-76ba-47ba-9bdb-a4a942dadb18-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.251377 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-scripts" (OuterVolumeSpecName: "scripts") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.262453 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.298160 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.320428 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-config-data" (OuterVolumeSpecName: "config-data") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.352325 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.352366 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.352379 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.388660 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10254294-76ba-47ba-9bdb-a4a942dadb18" (UID: "10254294-76ba-47ba-9bdb-a4a942dadb18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.455878 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10254294-76ba-47ba-9bdb-a4a942dadb18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.502209 4758 generic.go:334] "Generic (PLEG): container finished" podID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerID="2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5" exitCode=0 Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.502420 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.502365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerDied","Data":"2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5"} Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.502529 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10254294-76ba-47ba-9bdb-a4a942dadb18","Type":"ContainerDied","Data":"4759e581bae15fe3b6931e7a6a45784f38e2bddc9c95cb0df240e8a07b73d1d5"} Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.502569 4758 scope.go:117] "RemoveContainer" containerID="295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.530061 4758 scope.go:117] "RemoveContainer" containerID="aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.590215 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.602178 4758 scope.go:117] "RemoveContainer" containerID="74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.614057 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.638439 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.639631 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="proxy-httpd" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.639668 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="proxy-httpd" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.639708 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-notification-agent" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.639718 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-notification-agent" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.639744 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="sg-core" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.639754 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="sg-core" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.639797 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-central-agent" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.639807 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-central-agent" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.640115 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-notification-agent" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.640152 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="ceilometer-central-agent" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.640171 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="sg-core" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.640187 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" containerName="proxy-httpd" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.645541 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.652411 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.652744 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.652951 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.653483 4758 scope.go:117] "RemoveContainer" containerID="2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.691715 4758 scope.go:117] "RemoveContainer" containerID="295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.692298 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35\": container with ID starting with 295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35 not found: ID does not exist" containerID="295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.692513 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35"} err="failed to get container status \"295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35\": rpc error: code = NotFound desc = could not find container \"295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35\": container with ID starting with 295700d3d99442864205cbfa7443be1bdb7c3cd663dd7954e0d8060722e00d35 not found: ID does not exist" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.692555 4758 scope.go:117] "RemoveContainer" containerID="aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.693546 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f\": container with ID starting with aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f not found: ID does not exist" containerID="aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.693613 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f"} err="failed to get container status \"aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f\": rpc error: code = NotFound desc = could not find container \"aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f\": container with ID starting with aa56ef4b95a2f7de70b2d900e4cca49b23df8e96eb7d52a728884fb9cfc35c8f not found: ID does not exist" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.693634 4758 scope.go:117] "RemoveContainer" containerID="74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.694458 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8\": container with ID starting with 74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8 not found: ID does not exist" containerID="74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.694493 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8"} err="failed to get container status \"74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8\": rpc error: code = NotFound desc = could not find container \"74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8\": container with ID starting with 74f1648772ccfdcaeef07beb765c008f9779eab06990dfaf3196308a822b24f8 not found: ID does not exist" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.694541 4758 scope.go:117] "RemoveContainer" containerID="2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5" Dec 03 17:20:15 crc kubenswrapper[4758]: E1203 17:20:15.695531 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5\": container with ID starting with 2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5 not found: ID does not exist" containerID="2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.695601 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5"} err="failed to get container status \"2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5\": rpc error: code = NotFound desc = could not find container \"2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5\": container with ID starting with 2b6b49d59d64bd3e14f2aeb70d775f5a07ca930ffa1042bbd8a7d6c22de08da5 not found: ID does not exist" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765603 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765712 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765794 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-scripts\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765858 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-log-httpd\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765897 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-config-data\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765948 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28q99\" (UniqueName: \"kubernetes.io/projected/6a6a7745-6018-49a7-bc87-47204e670b35-kube-api-access-28q99\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.765998 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-run-httpd\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.833049 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9tms7"] Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868063 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-scripts\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868134 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-log-httpd\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868164 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-config-data\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868199 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28q99\" (UniqueName: \"kubernetes.io/projected/6a6a7745-6018-49a7-bc87-47204e670b35-kube-api-access-28q99\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868233 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-run-httpd\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.868323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.869334 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-log-httpd\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.869432 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-run-httpd\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.878101 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-config-data\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.879751 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.881262 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-scripts\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.884311 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.898356 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28q99\" (UniqueName: \"kubernetes.io/projected/6a6a7745-6018-49a7-bc87-47204e670b35-kube-api-access-28q99\") pod \"ceilometer-0\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " pod="openstack/ceilometer-0" Dec 03 17:20:15 crc kubenswrapper[4758]: I1203 17:20:15.973922 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:16 crc kubenswrapper[4758]: I1203 17:20:16.522735 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b49b1d3-cbb4-49da-8244-69e15595da33","Type":"ContainerStarted","Data":"b226eaf6e02f5bf23bb058d2357c7eb31da688995dc12dbdc9fe34fe8dbc7091"} Dec 03 17:20:16 crc kubenswrapper[4758]: I1203 17:20:16.528012 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9tms7" event={"ID":"52cd754b-1015-487b-a51c-4d7117cfae33","Type":"ContainerStarted","Data":"3fd39f4ffc9f6ed69ee9c24fa636e2459635be507e56b825b3b04b7ad56b51ef"} Dec 03 17:20:16 crc kubenswrapper[4758]: I1203 17:20:16.554468 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.554437371 podStartE2EDuration="4.554437371s" podCreationTimestamp="2025-12-03 17:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:20:16.544553666 +0000 UTC m=+1471.745930537" watchObservedRunningTime="2025-12-03 17:20:16.554437371 +0000 UTC m=+1471.755814232" Dec 03 17:20:16 crc kubenswrapper[4758]: W1203 17:20:16.558220 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a6a7745_6018_49a7_bc87_47204e670b35.slice/crio-90d9c6edfa79e9db71e74718fb4ed9efb69d3c0baf3a54b5987643e545529a8f WatchSource:0}: Error finding container 90d9c6edfa79e9db71e74718fb4ed9efb69d3c0baf3a54b5987643e545529a8f: Status 404 returned error can't find the container with id 90d9c6edfa79e9db71e74718fb4ed9efb69d3c0baf3a54b5987643e545529a8f Dec 03 17:20:16 crc kubenswrapper[4758]: I1203 17:20:16.570524 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:17 crc kubenswrapper[4758]: I1203 17:20:17.169399 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10254294-76ba-47ba-9bdb-a4a942dadb18" path="/var/lib/kubelet/pods/10254294-76ba-47ba-9bdb-a4a942dadb18/volumes" Dec 03 17:20:17 crc kubenswrapper[4758]: I1203 17:20:17.540077 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:17 crc kubenswrapper[4758]: I1203 17:20:17.560364 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerStarted","Data":"b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0"} Dec 03 17:20:17 crc kubenswrapper[4758]: I1203 17:20:17.560428 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerStarted","Data":"90d9c6edfa79e9db71e74718fb4ed9efb69d3c0baf3a54b5987643e545529a8f"} Dec 03 17:20:18 crc kubenswrapper[4758]: I1203 17:20:18.573335 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerStarted","Data":"05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7"} Dec 03 17:20:19 crc kubenswrapper[4758]: I1203 17:20:19.602079 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerStarted","Data":"cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d"} Dec 03 17:20:20 crc kubenswrapper[4758]: I1203 17:20:20.054981 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 17:20:20 crc kubenswrapper[4758]: I1203 17:20:20.057773 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 17:20:20 crc kubenswrapper[4758]: I1203 17:20:20.103877 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 17:20:20 crc kubenswrapper[4758]: I1203 17:20:20.132419 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 17:20:20 crc kubenswrapper[4758]: I1203 17:20:20.614425 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 17:20:20 crc kubenswrapper[4758]: I1203 17:20:20.614860 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 17:20:22 crc kubenswrapper[4758]: I1203 17:20:22.879295 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:22 crc kubenswrapper[4758]: I1203 17:20:22.879386 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:22 crc kubenswrapper[4758]: I1203 17:20:22.929477 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:22 crc kubenswrapper[4758]: I1203 17:20:22.964539 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:23 crc kubenswrapper[4758]: I1203 17:20:23.072928 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 17:20:23 crc kubenswrapper[4758]: I1203 17:20:23.073084 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:20:23 crc kubenswrapper[4758]: I1203 17:20:23.074414 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 17:20:23 crc kubenswrapper[4758]: I1203 17:20:23.676449 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:23 crc kubenswrapper[4758]: I1203 17:20:23.676508 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:25 crc kubenswrapper[4758]: I1203 17:20:25.497054 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9kgn"] Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.501323 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.519940 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9kgn"] Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.608739 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4h89\" (UniqueName: \"kubernetes.io/projected/bf5889fc-b255-421f-a14b-53a5bb4863a0-kube-api-access-d4h89\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.608831 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-utilities\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.609141 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-catalog-content\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.712093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4h89\" (UniqueName: \"kubernetes.io/projected/bf5889fc-b255-421f-a14b-53a5bb4863a0-kube-api-access-d4h89\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.712176 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-utilities\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.712229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-catalog-content\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.712849 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-utilities\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.712887 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-catalog-content\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.738812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4h89\" (UniqueName: \"kubernetes.io/projected/bf5889fc-b255-421f-a14b-53a5bb4863a0-kube-api-access-d4h89\") pod \"redhat-operators-l9kgn\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:25.829128 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:26.114658 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:26.114780 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:20:26 crc kubenswrapper[4758]: I1203 17:20:26.295304 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 17:20:30 crc kubenswrapper[4758]: E1203 17:20:30.400978 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Dec 03 17:20:30 crc kubenswrapper[4758]: E1203 17:20:30.401969 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pk7cj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-9tms7_openstack(52cd754b-1015-487b-a51c-4d7117cfae33): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:20:30 crc kubenswrapper[4758]: E1203 17:20:30.403214 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-9tms7" podUID="52cd754b-1015-487b-a51c-4d7117cfae33" Dec 03 17:20:30 crc kubenswrapper[4758]: E1203 17:20:30.782697 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-9tms7" podUID="52cd754b-1015-487b-a51c-4d7117cfae33" Dec 03 17:20:30 crc kubenswrapper[4758]: I1203 17:20:30.955249 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9kgn"] Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.790519 4758 generic.go:334] "Generic (PLEG): container finished" podID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerID="1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c" exitCode=0 Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.790629 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerDied","Data":"1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c"} Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.790706 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerStarted","Data":"d0862f9dd239f79f306a4667a7099770e6616cf915d775f77e646711b75bebba"} Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.797119 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerStarted","Data":"bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128"} Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.797509 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.797311 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="sg-core" containerID="cri-o://cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d" gracePeriod=30 Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.797326 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-notification-agent" containerID="cri-o://05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7" gracePeriod=30 Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.797302 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="proxy-httpd" containerID="cri-o://bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128" gracePeriod=30 Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.797280 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-central-agent" containerID="cri-o://b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0" gracePeriod=30 Dec 03 17:20:31 crc kubenswrapper[4758]: I1203 17:20:31.864956 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.896833619 podStartE2EDuration="16.864929497s" podCreationTimestamp="2025-12-03 17:20:15 +0000 UTC" firstStartedPulling="2025-12-03 17:20:16.560940866 +0000 UTC m=+1471.762317727" lastFinishedPulling="2025-12-03 17:20:30.529036744 +0000 UTC m=+1485.730413605" observedRunningTime="2025-12-03 17:20:31.856655895 +0000 UTC m=+1487.058032776" watchObservedRunningTime="2025-12-03 17:20:31.864929497 +0000 UTC m=+1487.066306358" Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.814006 4758 generic.go:334] "Generic (PLEG): container finished" podID="6a6a7745-6018-49a7-bc87-47204e670b35" containerID="bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128" exitCode=0 Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.814465 4758 generic.go:334] "Generic (PLEG): container finished" podID="6a6a7745-6018-49a7-bc87-47204e670b35" containerID="cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d" exitCode=2 Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.814479 4758 generic.go:334] "Generic (PLEG): container finished" podID="6a6a7745-6018-49a7-bc87-47204e670b35" containerID="b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0" exitCode=0 Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.814078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerDied","Data":"bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128"} Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.814603 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerDied","Data":"cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d"} Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.814622 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerDied","Data":"b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0"} Dec 03 17:20:32 crc kubenswrapper[4758]: I1203 17:20:32.816890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerStarted","Data":"be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39"} Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.356713 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.505482 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-log-httpd\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.505549 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-run-httpd\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.505577 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-combined-ca-bundle\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.505636 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28q99\" (UniqueName: \"kubernetes.io/projected/6a6a7745-6018-49a7-bc87-47204e670b35-kube-api-access-28q99\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.505788 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-sg-core-conf-yaml\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.505810 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-scripts\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.506021 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-config-data\") pod \"6a6a7745-6018-49a7-bc87-47204e670b35\" (UID: \"6a6a7745-6018-49a7-bc87-47204e670b35\") " Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.507886 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.509497 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.513585 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a6a7745-6018-49a7-bc87-47204e670b35-kube-api-access-28q99" (OuterVolumeSpecName: "kube-api-access-28q99") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "kube-api-access-28q99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.516631 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-scripts" (OuterVolumeSpecName: "scripts") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.544624 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.596375 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.609360 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.609410 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a6a7745-6018-49a7-bc87-47204e670b35-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.609425 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.609439 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28q99\" (UniqueName: \"kubernetes.io/projected/6a6a7745-6018-49a7-bc87-47204e670b35-kube-api-access-28q99\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.609480 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.609495 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.645909 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-config-data" (OuterVolumeSpecName: "config-data") pod "6a6a7745-6018-49a7-bc87-47204e670b35" (UID: "6a6a7745-6018-49a7-bc87-47204e670b35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.712086 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a6a7745-6018-49a7-bc87-47204e670b35-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.833001 4758 generic.go:334] "Generic (PLEG): container finished" podID="6a6a7745-6018-49a7-bc87-47204e670b35" containerID="05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7" exitCode=0 Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.833094 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerDied","Data":"05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7"} Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.833132 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a6a7745-6018-49a7-bc87-47204e670b35","Type":"ContainerDied","Data":"90d9c6edfa79e9db71e74718fb4ed9efb69d3c0baf3a54b5987643e545529a8f"} Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.833157 4758 scope.go:117] "RemoveContainer" containerID="bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.833343 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.846477 4758 generic.go:334] "Generic (PLEG): container finished" podID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerID="be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39" exitCode=0 Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.846549 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerDied","Data":"be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39"} Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.872675 4758 scope.go:117] "RemoveContainer" containerID="cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.905536 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.908664 4758 scope.go:117] "RemoveContainer" containerID="05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.919220 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.946454 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:33 crc kubenswrapper[4758]: E1203 17:20:33.947175 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-notification-agent" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947205 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-notification-agent" Dec 03 17:20:33 crc kubenswrapper[4758]: E1203 17:20:33.947247 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-central-agent" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947275 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-central-agent" Dec 03 17:20:33 crc kubenswrapper[4758]: E1203 17:20:33.947298 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="sg-core" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947307 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="sg-core" Dec 03 17:20:33 crc kubenswrapper[4758]: E1203 17:20:33.947330 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="proxy-httpd" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947339 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="proxy-httpd" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947568 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-notification-agent" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947617 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="proxy-httpd" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947628 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="sg-core" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.947642 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" containerName="ceilometer-central-agent" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.950237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.954097 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.954392 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.958583 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:33 crc kubenswrapper[4758]: I1203 17:20:33.967174 4758 scope.go:117] "RemoveContainer" containerID="b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.019183 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.019265 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-log-httpd\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.019291 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-config-data\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.019929 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.020017 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9jz4\" (UniqueName: \"kubernetes.io/projected/c8850265-8985-4526-9143-66966b163609-kube-api-access-b9jz4\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.020175 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-run-httpd\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.020303 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-scripts\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.023277 4758 scope.go:117] "RemoveContainer" containerID="bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128" Dec 03 17:20:34 crc kubenswrapper[4758]: E1203 17:20:34.024348 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128\": container with ID starting with bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128 not found: ID does not exist" containerID="bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.024380 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128"} err="failed to get container status \"bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128\": rpc error: code = NotFound desc = could not find container \"bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128\": container with ID starting with bcf84ae426620772979e55d8ca0892e410fd389c379d7c65538806f27a0c1128 not found: ID does not exist" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.024406 4758 scope.go:117] "RemoveContainer" containerID="cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d" Dec 03 17:20:34 crc kubenswrapper[4758]: E1203 17:20:34.025241 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d\": container with ID starting with cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d not found: ID does not exist" containerID="cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.025266 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d"} err="failed to get container status \"cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d\": rpc error: code = NotFound desc = could not find container \"cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d\": container with ID starting with cc4027c45a55a41942925c65269a1284ed87360ba667d69661cd0baa3c15570d not found: ID does not exist" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.025289 4758 scope.go:117] "RemoveContainer" containerID="05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7" Dec 03 17:20:34 crc kubenswrapper[4758]: E1203 17:20:34.025577 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7\": container with ID starting with 05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7 not found: ID does not exist" containerID="05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.025595 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7"} err="failed to get container status \"05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7\": rpc error: code = NotFound desc = could not find container \"05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7\": container with ID starting with 05b86578783fe33bd06cbf106124c92619bc82233239fa263b7dfa5766475ce7 not found: ID does not exist" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.025641 4758 scope.go:117] "RemoveContainer" containerID="b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0" Dec 03 17:20:34 crc kubenswrapper[4758]: E1203 17:20:34.026663 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0\": container with ID starting with b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0 not found: ID does not exist" containerID="b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.026702 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0"} err="failed to get container status \"b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0\": rpc error: code = NotFound desc = could not find container \"b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0\": container with ID starting with b26a215d777cb64ea80247b85a35a694c05281653afd737290f50bf6770e57e0 not found: ID does not exist" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.125463 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-scripts\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.125850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.125913 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-log-httpd\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.125974 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-config-data\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.126033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.126075 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9jz4\" (UniqueName: \"kubernetes.io/projected/c8850265-8985-4526-9143-66966b163609-kube-api-access-b9jz4\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.126155 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-run-httpd\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.126879 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-run-httpd\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.132702 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-scripts\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.133087 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-log-httpd\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.136781 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-config-data\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.139461 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:34 crc kubenswrapper[4758]: I1203 17:20:34.140110 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:35 crc kubenswrapper[4758]: I1203 17:20:35.129577 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a6a7745-6018-49a7-bc87-47204e670b35" path="/var/lib/kubelet/pods/6a6a7745-6018-49a7-bc87-47204e670b35/volumes" Dec 03 17:20:35 crc kubenswrapper[4758]: I1203 17:20:35.282094 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9jz4\" (UniqueName: \"kubernetes.io/projected/c8850265-8985-4526-9143-66966b163609-kube-api-access-b9jz4\") pod \"ceilometer-0\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " pod="openstack/ceilometer-0" Dec 03 17:20:35 crc kubenswrapper[4758]: I1203 17:20:35.487200 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:35 crc kubenswrapper[4758]: I1203 17:20:35.640215 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:35 crc kubenswrapper[4758]: I1203 17:20:35.891120 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerStarted","Data":"b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b"} Dec 03 17:20:35 crc kubenswrapper[4758]: I1203 17:20:35.932341 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9kgn" podStartSLOduration=7.375496988 podStartE2EDuration="10.932302644s" podCreationTimestamp="2025-12-03 17:20:25 +0000 UTC" firstStartedPulling="2025-12-03 17:20:31.79349393 +0000 UTC m=+1486.994870791" lastFinishedPulling="2025-12-03 17:20:35.350299586 +0000 UTC m=+1490.551676447" observedRunningTime="2025-12-03 17:20:35.921330169 +0000 UTC m=+1491.122707040" watchObservedRunningTime="2025-12-03 17:20:35.932302644 +0000 UTC m=+1491.133679505" Dec 03 17:20:36 crc kubenswrapper[4758]: I1203 17:20:36.132139 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:36 crc kubenswrapper[4758]: I1203 17:20:36.907003 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerStarted","Data":"1be10e2bea0e92f1040ce2acbcc1a10cf581876e680a1812f3732069264aada6"} Dec 03 17:20:38 crc kubenswrapper[4758]: I1203 17:20:38.932196 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerStarted","Data":"1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4"} Dec 03 17:20:38 crc kubenswrapper[4758]: I1203 17:20:38.933098 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerStarted","Data":"9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376"} Dec 03 17:20:39 crc kubenswrapper[4758]: I1203 17:20:39.952521 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerStarted","Data":"6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094"} Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.515720 4758 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podb8eb3ef8-1b17-4d2d-b664-f32d7263d559"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podb8eb3ef8-1b17-4d2d-b664-f32d7263d559] : Timed out while waiting for systemd to remove kubepods-besteffort-podb8eb3ef8_1b17_4d2d_b664_f32d7263d559.slice" Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.983783 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerStarted","Data":"becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8"} Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.984022 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-central-agent" containerID="cri-o://9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376" gracePeriod=30 Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.984178 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-notification-agent" containerID="cri-o://1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4" gracePeriod=30 Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.984274 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8850265-8985-4526-9143-66966b163609" containerName="sg-core" containerID="cri-o://6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094" gracePeriod=30 Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.984187 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8850265-8985-4526-9143-66966b163609" containerName="proxy-httpd" containerID="cri-o://becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8" gracePeriod=30 Dec 03 17:20:41 crc kubenswrapper[4758]: I1203 17:20:41.984436 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:20:42 crc kubenswrapper[4758]: I1203 17:20:42.011836 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.925136826 podStartE2EDuration="9.011813264s" podCreationTimestamp="2025-12-03 17:20:33 +0000 UTC" firstStartedPulling="2025-12-03 17:20:36.152457611 +0000 UTC m=+1491.353834472" lastFinishedPulling="2025-12-03 17:20:41.239134049 +0000 UTC m=+1496.440510910" observedRunningTime="2025-12-03 17:20:42.009541462 +0000 UTC m=+1497.210918323" watchObservedRunningTime="2025-12-03 17:20:42.011813264 +0000 UTC m=+1497.213190125" Dec 03 17:20:43 crc kubenswrapper[4758]: I1203 17:20:43.001391 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerDied","Data":"becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8"} Dec 03 17:20:43 crc kubenswrapper[4758]: I1203 17:20:43.001350 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8850265-8985-4526-9143-66966b163609" containerID="becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8" exitCode=0 Dec 03 17:20:43 crc kubenswrapper[4758]: I1203 17:20:43.001500 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8850265-8985-4526-9143-66966b163609" containerID="6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094" exitCode=2 Dec 03 17:20:43 crc kubenswrapper[4758]: I1203 17:20:43.001522 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8850265-8985-4526-9143-66966b163609" containerID="1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4" exitCode=0 Dec 03 17:20:43 crc kubenswrapper[4758]: I1203 17:20:43.001542 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerDied","Data":"6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094"} Dec 03 17:20:43 crc kubenswrapper[4758]: I1203 17:20:43.001556 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerDied","Data":"1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4"} Dec 03 17:20:44 crc kubenswrapper[4758]: I1203 17:20:44.014543 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9tms7" event={"ID":"52cd754b-1015-487b-a51c-4d7117cfae33","Type":"ContainerStarted","Data":"8a484e9fea768918ed84dcc3fd4d0716456f1c33003094d4dfccc9c14d6684a7"} Dec 03 17:20:44 crc kubenswrapper[4758]: I1203 17:20:44.035271 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-9tms7" podStartSLOduration=2.84555465 podStartE2EDuration="30.035249341s" podCreationTimestamp="2025-12-03 17:20:14 +0000 UTC" firstStartedPulling="2025-12-03 17:20:15.83526681 +0000 UTC m=+1471.036643671" lastFinishedPulling="2025-12-03 17:20:43.024961501 +0000 UTC m=+1498.226338362" observedRunningTime="2025-12-03 17:20:44.032329423 +0000 UTC m=+1499.233706294" watchObservedRunningTime="2025-12-03 17:20:44.035249341 +0000 UTC m=+1499.236626202" Dec 03 17:20:45 crc kubenswrapper[4758]: I1203 17:20:45.829546 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:45 crc kubenswrapper[4758]: I1203 17:20:45.830361 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:46 crc kubenswrapper[4758]: I1203 17:20:46.887122 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9kgn" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="registry-server" probeResult="failure" output=< Dec 03 17:20:46 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 17:20:46 crc kubenswrapper[4758]: > Dec 03 17:20:48 crc kubenswrapper[4758]: I1203 17:20:48.976647 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.083637 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-log-httpd\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.083858 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-config-data\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.083964 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-scripts\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.084003 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-combined-ca-bundle\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.084104 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-run-httpd\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.084139 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9jz4\" (UniqueName: \"kubernetes.io/projected/c8850265-8985-4526-9143-66966b163609-kube-api-access-b9jz4\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.084178 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-sg-core-conf-yaml\") pod \"c8850265-8985-4526-9143-66966b163609\" (UID: \"c8850265-8985-4526-9143-66966b163609\") " Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.084309 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.084722 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.086952 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.103074 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8850265-8985-4526-9143-66966b163609-kube-api-access-b9jz4" (OuterVolumeSpecName: "kube-api-access-b9jz4") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "kube-api-access-b9jz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.111990 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8850265-8985-4526-9143-66966b163609" containerID="9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376" exitCode=0 Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.112054 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerDied","Data":"9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376"} Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.112097 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8850265-8985-4526-9143-66966b163609","Type":"ContainerDied","Data":"1be10e2bea0e92f1040ce2acbcc1a10cf581876e680a1812f3732069264aada6"} Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.112121 4758 scope.go:117] "RemoveContainer" containerID="becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.112333 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.127601 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-scripts" (OuterVolumeSpecName: "scripts") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.162150 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.188593 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.188635 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8850265-8985-4526-9143-66966b163609-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.188646 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9jz4\" (UniqueName: \"kubernetes.io/projected/c8850265-8985-4526-9143-66966b163609-kube-api-access-b9jz4\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.188657 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.253826 4758 scope.go:117] "RemoveContainer" containerID="6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.271718 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.291620 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.305513 4758 scope.go:117] "RemoveContainer" containerID="1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.339070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-config-data" (OuterVolumeSpecName: "config-data") pod "c8850265-8985-4526-9143-66966b163609" (UID: "c8850265-8985-4526-9143-66966b163609"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.339559 4758 scope.go:117] "RemoveContainer" containerID="9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.365897 4758 scope.go:117] "RemoveContainer" containerID="becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.366608 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8\": container with ID starting with becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8 not found: ID does not exist" containerID="becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.366853 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8"} err="failed to get container status \"becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8\": rpc error: code = NotFound desc = could not find container \"becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8\": container with ID starting with becf668e2b8324557f2b3fcaf8bf189433788d93ee62f6d62430e1689163e5e8 not found: ID does not exist" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.366910 4758 scope.go:117] "RemoveContainer" containerID="6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.367425 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094\": container with ID starting with 6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094 not found: ID does not exist" containerID="6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.367477 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094"} err="failed to get container status \"6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094\": rpc error: code = NotFound desc = could not find container \"6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094\": container with ID starting with 6bf6228718c4061ed84eeac51a9d82732c0734fa9b4e44c045ee782cac7cb094 not found: ID does not exist" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.367511 4758 scope.go:117] "RemoveContainer" containerID="1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.367888 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4\": container with ID starting with 1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4 not found: ID does not exist" containerID="1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.367953 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4"} err="failed to get container status \"1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4\": rpc error: code = NotFound desc = could not find container \"1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4\": container with ID starting with 1dc08cb4da2ba022844bdd1aa69be117665f9454d1646bca079f8048f9aa51f4 not found: ID does not exist" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.367974 4758 scope.go:117] "RemoveContainer" containerID="9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.368506 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376\": container with ID starting with 9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376 not found: ID does not exist" containerID="9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.368569 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376"} err="failed to get container status \"9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376\": rpc error: code = NotFound desc = could not find container \"9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376\": container with ID starting with 9a86480889a7a50274201c1d7ea788f45c0c5af3275e606fa92e6d08170a7376 not found: ID does not exist" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.394240 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8850265-8985-4526-9143-66966b163609-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.456884 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.467552 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.499370 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.499878 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8850265-8985-4526-9143-66966b163609" containerName="sg-core" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.499898 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8850265-8985-4526-9143-66966b163609" containerName="sg-core" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.499926 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-notification-agent" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.499934 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-notification-agent" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.499949 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8850265-8985-4526-9143-66966b163609" containerName="proxy-httpd" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.499955 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8850265-8985-4526-9143-66966b163609" containerName="proxy-httpd" Dec 03 17:20:49 crc kubenswrapper[4758]: E1203 17:20:49.499972 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-central-agent" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.499978 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-central-agent" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.500173 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-central-agent" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.500188 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8850265-8985-4526-9143-66966b163609" containerName="sg-core" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.500207 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8850265-8985-4526-9143-66966b163609" containerName="proxy-httpd" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.500221 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8850265-8985-4526-9143-66966b163609" containerName="ceilometer-notification-agent" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.501976 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.506515 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.507586 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.518753 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.707609 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-run-httpd\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.707697 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-log-httpd\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.707735 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.707782 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-scripts\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.708301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-config-data\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.708436 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.708520 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2dzl\" (UniqueName: \"kubernetes.io/projected/03107cd1-508b-404f-9728-43893717064b-kube-api-access-g2dzl\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811059 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811128 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2dzl\" (UniqueName: \"kubernetes.io/projected/03107cd1-508b-404f-9728-43893717064b-kube-api-access-g2dzl\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-run-httpd\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811255 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-log-httpd\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811334 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-scripts\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.811411 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-config-data\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.812280 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-log-httpd\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.813432 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-run-httpd\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.819034 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-scripts\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.819054 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.819123 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.819979 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-config-data\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.838630 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2dzl\" (UniqueName: \"kubernetes.io/projected/03107cd1-508b-404f-9728-43893717064b-kube-api-access-g2dzl\") pod \"ceilometer-0\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " pod="openstack/ceilometer-0" Dec 03 17:20:49 crc kubenswrapper[4758]: I1203 17:20:49.839632 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:20:50 crc kubenswrapper[4758]: I1203 17:20:50.348471 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:20:51 crc kubenswrapper[4758]: I1203 17:20:51.130349 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8850265-8985-4526-9143-66966b163609" path="/var/lib/kubelet/pods/c8850265-8985-4526-9143-66966b163609/volumes" Dec 03 17:20:51 crc kubenswrapper[4758]: I1203 17:20:51.143647 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerStarted","Data":"15876892b9da8a327b729f697291d12ccc619a91e6d8d565a0aa68c167ebe668"} Dec 03 17:20:52 crc kubenswrapper[4758]: I1203 17:20:52.157312 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerStarted","Data":"5d5d520d0a731b3f0bdb5b1c023550ac627838a6ed151ddec281961c090e9500"} Dec 03 17:20:52 crc kubenswrapper[4758]: I1203 17:20:52.157770 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerStarted","Data":"bba07a358e536f9d95a161647edd4d6a2ad3bfc0d86d90e9cc74f3fe0291df0b"} Dec 03 17:20:53 crc kubenswrapper[4758]: I1203 17:20:53.178609 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerStarted","Data":"7258d5fbd64e678990cf01affedc429932e83f78180f40d227f66e10a4188120"} Dec 03 17:20:55 crc kubenswrapper[4758]: I1203 17:20:55.205621 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerStarted","Data":"c7c26dd0693f5732f58928023400c924fb34a8c2dcf2d6d240a6a91560f10d08"} Dec 03 17:20:55 crc kubenswrapper[4758]: I1203 17:20:55.206149 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:20:55 crc kubenswrapper[4758]: I1203 17:20:55.233160 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.454654597 podStartE2EDuration="6.23313007s" podCreationTimestamp="2025-12-03 17:20:49 +0000 UTC" firstStartedPulling="2025-12-03 17:20:50.352403519 +0000 UTC m=+1505.553780380" lastFinishedPulling="2025-12-03 17:20:54.130878982 +0000 UTC m=+1509.332255853" observedRunningTime="2025-12-03 17:20:55.231025453 +0000 UTC m=+1510.432402314" watchObservedRunningTime="2025-12-03 17:20:55.23313007 +0000 UTC m=+1510.434506931" Dec 03 17:20:55 crc kubenswrapper[4758]: I1203 17:20:55.890301 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:55 crc kubenswrapper[4758]: I1203 17:20:55.942558 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:56 crc kubenswrapper[4758]: I1203 17:20:56.710371 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9kgn"] Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.240602 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9kgn" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="registry-server" containerID="cri-o://b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b" gracePeriod=2 Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.777377 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.930400 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-catalog-content\") pod \"bf5889fc-b255-421f-a14b-53a5bb4863a0\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.930495 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-utilities\") pod \"bf5889fc-b255-421f-a14b-53a5bb4863a0\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.930565 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4h89\" (UniqueName: \"kubernetes.io/projected/bf5889fc-b255-421f-a14b-53a5bb4863a0-kube-api-access-d4h89\") pod \"bf5889fc-b255-421f-a14b-53a5bb4863a0\" (UID: \"bf5889fc-b255-421f-a14b-53a5bb4863a0\") " Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.931701 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-utilities" (OuterVolumeSpecName: "utilities") pod "bf5889fc-b255-421f-a14b-53a5bb4863a0" (UID: "bf5889fc-b255-421f-a14b-53a5bb4863a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:57 crc kubenswrapper[4758]: I1203 17:20:57.937902 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf5889fc-b255-421f-a14b-53a5bb4863a0-kube-api-access-d4h89" (OuterVolumeSpecName: "kube-api-access-d4h89") pod "bf5889fc-b255-421f-a14b-53a5bb4863a0" (UID: "bf5889fc-b255-421f-a14b-53a5bb4863a0"). InnerVolumeSpecName "kube-api-access-d4h89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.033231 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.033287 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4h89\" (UniqueName: \"kubernetes.io/projected/bf5889fc-b255-421f-a14b-53a5bb4863a0-kube-api-access-d4h89\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.045957 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf5889fc-b255-421f-a14b-53a5bb4863a0" (UID: "bf5889fc-b255-421f-a14b-53a5bb4863a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.135487 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5889fc-b255-421f-a14b-53a5bb4863a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.254834 4758 generic.go:334] "Generic (PLEG): container finished" podID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerID="b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b" exitCode=0 Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.254940 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9kgn" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.254969 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerDied","Data":"b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b"} Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.256971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9kgn" event={"ID":"bf5889fc-b255-421f-a14b-53a5bb4863a0","Type":"ContainerDied","Data":"d0862f9dd239f79f306a4667a7099770e6616cf915d775f77e646711b75bebba"} Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.256997 4758 scope.go:117] "RemoveContainer" containerID="b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.295256 4758 scope.go:117] "RemoveContainer" containerID="be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.301534 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9kgn"] Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.312909 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9kgn"] Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.323121 4758 scope.go:117] "RemoveContainer" containerID="1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.372339 4758 scope.go:117] "RemoveContainer" containerID="b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b" Dec 03 17:20:58 crc kubenswrapper[4758]: E1203 17:20:58.373013 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b\": container with ID starting with b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b not found: ID does not exist" containerID="b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.373072 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b"} err="failed to get container status \"b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b\": rpc error: code = NotFound desc = could not find container \"b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b\": container with ID starting with b3b219f0bea50c6af1ef74e7b401e2d0956e9e4f0d07a2f0b0b822e9765dd65b not found: ID does not exist" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.373118 4758 scope.go:117] "RemoveContainer" containerID="be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39" Dec 03 17:20:58 crc kubenswrapper[4758]: E1203 17:20:58.373467 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39\": container with ID starting with be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39 not found: ID does not exist" containerID="be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.373502 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39"} err="failed to get container status \"be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39\": rpc error: code = NotFound desc = could not find container \"be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39\": container with ID starting with be4e056a4d1ae328124935b25b6672ad4c0506ff0b42b92d3d02045858941b39 not found: ID does not exist" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.373535 4758 scope.go:117] "RemoveContainer" containerID="1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c" Dec 03 17:20:58 crc kubenswrapper[4758]: E1203 17:20:58.373982 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c\": container with ID starting with 1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c not found: ID does not exist" containerID="1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c" Dec 03 17:20:58 crc kubenswrapper[4758]: I1203 17:20:58.374047 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c"} err="failed to get container status \"1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c\": rpc error: code = NotFound desc = could not find container \"1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c\": container with ID starting with 1877b8c9161d35ef19b073cc8a406de5932548b9ffcebac4de6c155ea825418c not found: ID does not exist" Dec 03 17:20:59 crc kubenswrapper[4758]: I1203 17:20:59.128726 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" path="/var/lib/kubelet/pods/bf5889fc-b255-421f-a14b-53a5bb4863a0/volumes" Dec 03 17:21:00 crc kubenswrapper[4758]: I1203 17:21:00.283745 4758 generic.go:334] "Generic (PLEG): container finished" podID="52cd754b-1015-487b-a51c-4d7117cfae33" containerID="8a484e9fea768918ed84dcc3fd4d0716456f1c33003094d4dfccc9c14d6684a7" exitCode=0 Dec 03 17:21:00 crc kubenswrapper[4758]: I1203 17:21:00.283819 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9tms7" event={"ID":"52cd754b-1015-487b-a51c-4d7117cfae33","Type":"ContainerDied","Data":"8a484e9fea768918ed84dcc3fd4d0716456f1c33003094d4dfccc9c14d6684a7"} Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.686246 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.822384 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-config-data\") pod \"52cd754b-1015-487b-a51c-4d7117cfae33\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.822519 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk7cj\" (UniqueName: \"kubernetes.io/projected/52cd754b-1015-487b-a51c-4d7117cfae33-kube-api-access-pk7cj\") pod \"52cd754b-1015-487b-a51c-4d7117cfae33\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.822572 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-scripts\") pod \"52cd754b-1015-487b-a51c-4d7117cfae33\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.822753 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-combined-ca-bundle\") pod \"52cd754b-1015-487b-a51c-4d7117cfae33\" (UID: \"52cd754b-1015-487b-a51c-4d7117cfae33\") " Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.831000 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-scripts" (OuterVolumeSpecName: "scripts") pod "52cd754b-1015-487b-a51c-4d7117cfae33" (UID: "52cd754b-1015-487b-a51c-4d7117cfae33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.831272 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cd754b-1015-487b-a51c-4d7117cfae33-kube-api-access-pk7cj" (OuterVolumeSpecName: "kube-api-access-pk7cj") pod "52cd754b-1015-487b-a51c-4d7117cfae33" (UID: "52cd754b-1015-487b-a51c-4d7117cfae33"). InnerVolumeSpecName "kube-api-access-pk7cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.861238 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52cd754b-1015-487b-a51c-4d7117cfae33" (UID: "52cd754b-1015-487b-a51c-4d7117cfae33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.863212 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-config-data" (OuterVolumeSpecName: "config-data") pod "52cd754b-1015-487b-a51c-4d7117cfae33" (UID: "52cd754b-1015-487b-a51c-4d7117cfae33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.926783 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.926834 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk7cj\" (UniqueName: \"kubernetes.io/projected/52cd754b-1015-487b-a51c-4d7117cfae33-kube-api-access-pk7cj\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.926850 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:01 crc kubenswrapper[4758]: I1203 17:21:01.926859 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52cd754b-1015-487b-a51c-4d7117cfae33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.309236 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9tms7" event={"ID":"52cd754b-1015-487b-a51c-4d7117cfae33","Type":"ContainerDied","Data":"3fd39f4ffc9f6ed69ee9c24fa636e2459635be507e56b825b3b04b7ad56b51ef"} Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.309821 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fd39f4ffc9f6ed69ee9c24fa636e2459635be507e56b825b3b04b7ad56b51ef" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.309382 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9tms7" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.493356 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:21:02 crc kubenswrapper[4758]: E1203 17:21:02.493925 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52cd754b-1015-487b-a51c-4d7117cfae33" containerName="nova-cell0-conductor-db-sync" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.493946 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="52cd754b-1015-487b-a51c-4d7117cfae33" containerName="nova-cell0-conductor-db-sync" Dec 03 17:21:02 crc kubenswrapper[4758]: E1203 17:21:02.493974 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="extract-utilities" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.493982 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="extract-utilities" Dec 03 17:21:02 crc kubenswrapper[4758]: E1203 17:21:02.493992 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="registry-server" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.494003 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="registry-server" Dec 03 17:21:02 crc kubenswrapper[4758]: E1203 17:21:02.494022 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="extract-content" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.494028 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="extract-content" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.494204 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf5889fc-b255-421f-a14b-53a5bb4863a0" containerName="registry-server" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.494218 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="52cd754b-1015-487b-a51c-4d7117cfae33" containerName="nova-cell0-conductor-db-sync" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.495955 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.501274 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6fv2h" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.501332 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.539028 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.643525 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.643643 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.643830 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sw8d\" (UniqueName: \"kubernetes.io/projected/ebee961b-48e5-4983-bb45-c14bbcc164f8-kube-api-access-8sw8d\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.746046 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sw8d\" (UniqueName: \"kubernetes.io/projected/ebee961b-48e5-4983-bb45-c14bbcc164f8-kube-api-access-8sw8d\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.746254 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.746378 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.753113 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.753386 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.770104 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sw8d\" (UniqueName: \"kubernetes.io/projected/ebee961b-48e5-4983-bb45-c14bbcc164f8-kube-api-access-8sw8d\") pod \"nova-cell0-conductor-0\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:02 crc kubenswrapper[4758]: I1203 17:21:02.832718 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:03 crc kubenswrapper[4758]: I1203 17:21:03.293142 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:21:03 crc kubenswrapper[4758]: I1203 17:21:03.333806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebee961b-48e5-4983-bb45-c14bbcc164f8","Type":"ContainerStarted","Data":"8f73a844ae769c58acc396ebbb266655e2232b126e23cdb30ff2999baaaabeca"} Dec 03 17:21:04 crc kubenswrapper[4758]: I1203 17:21:04.350498 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebee961b-48e5-4983-bb45-c14bbcc164f8","Type":"ContainerStarted","Data":"b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75"} Dec 03 17:21:04 crc kubenswrapper[4758]: I1203 17:21:04.351069 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:04 crc kubenswrapper[4758]: I1203 17:21:04.380591 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.380560415 podStartE2EDuration="2.380560415s" podCreationTimestamp="2025-12-03 17:21:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:04.373128396 +0000 UTC m=+1519.574505267" watchObservedRunningTime="2025-12-03 17:21:04.380560415 +0000 UTC m=+1519.581937276" Dec 03 17:21:12 crc kubenswrapper[4758]: I1203 17:21:12.861885 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.391864 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-w5n6b"] Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.393503 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.399324 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.399731 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.410028 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w5n6b"] Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.488021 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv4nt\" (UniqueName: \"kubernetes.io/projected/efe2b439-b61f-4a3e-accc-abc261fc216f-kube-api-access-xv4nt\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.488102 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.488138 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-config-data\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.488195 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-scripts\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.594492 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.594582 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-config-data\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.594641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-scripts\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.594779 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv4nt\" (UniqueName: \"kubernetes.io/projected/efe2b439-b61f-4a3e-accc-abc261fc216f-kube-api-access-xv4nt\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.607731 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.613397 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-scripts\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.630785 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-config-data\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.634811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv4nt\" (UniqueName: \"kubernetes.io/projected/efe2b439-b61f-4a3e-accc-abc261fc216f-kube-api-access-xv4nt\") pod \"nova-cell0-cell-mapping-w5n6b\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.723931 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.725282 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.726645 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.742355 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.779551 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.802174 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.802310 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k75kz\" (UniqueName: \"kubernetes.io/projected/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-kube-api-access-k75kz\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.802364 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.905476 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k75kz\" (UniqueName: \"kubernetes.io/projected/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-kube-api-access-k75kz\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.905592 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.905753 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.939498 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.940258 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:13 crc kubenswrapper[4758]: I1203 17:21:13.971584 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.015848 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k75kz\" (UniqueName: \"kubernetes.io/projected/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-kube-api-access-k75kz\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.070371 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.070599 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.095768 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.097173 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.163944 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.176107 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.176545 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.179770 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.202987 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.204998 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.210832 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.210989 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa876ee7-a6e5-4393-a387-a965eefb43dd-logs\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.211093 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-config-data\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.211346 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq78b\" (UniqueName: \"kubernetes.io/projected/aa876ee7-a6e5-4393-a387-a965eefb43dd-kube-api-access-vq78b\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.212395 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.215933 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.269122 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-f9t2n"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.271287 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.286518 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-f9t2n"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.313744 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317059 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c452f\" (UniqueName: \"kubernetes.io/projected/8e3486fc-aed7-4e82-8ad4-73b465917bbf-kube-api-access-c452f\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317352 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317530 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq78b\" (UniqueName: \"kubernetes.io/projected/aa876ee7-a6e5-4393-a387-a965eefb43dd-kube-api-access-vq78b\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317647 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317701 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3486fc-aed7-4e82-8ad4-73b465917bbf-logs\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317839 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cgsx\" (UniqueName: \"kubernetes.io/projected/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-kube-api-access-8cgsx\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-config-data\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317909 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa876ee7-a6e5-4393-a387-a965eefb43dd-logs\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.317955 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-config-data\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.318096 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-config-data\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.325102 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.327444 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa876ee7-a6e5-4393-a387-a965eefb43dd-logs\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.328747 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-config-data\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.341437 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq78b\" (UniqueName: \"kubernetes.io/projected/aa876ee7-a6e5-4393-a387-a965eefb43dd-kube-api-access-vq78b\") pod \"nova-api-0\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420263 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-svc\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420342 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3486fc-aed7-4e82-8ad4-73b465917bbf-logs\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420393 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzg6f\" (UniqueName: \"kubernetes.io/projected/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-kube-api-access-nzg6f\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-config-data\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420476 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cgsx\" (UniqueName: \"kubernetes.io/projected/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-kube-api-access-8cgsx\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420508 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-config-data\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420556 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420615 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420636 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-config\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.420657 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.421022 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c452f\" (UniqueName: \"kubernetes.io/projected/8e3486fc-aed7-4e82-8ad4-73b465917bbf-kube-api-access-c452f\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.421074 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.421109 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.424663 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3486fc-aed7-4e82-8ad4-73b465917bbf-logs\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.426271 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-config-data\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.431660 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.432213 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.433672 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-config-data\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.446488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cgsx\" (UniqueName: \"kubernetes.io/projected/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-kube-api-access-8cgsx\") pod \"nova-scheduler-0\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.462263 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c452f\" (UniqueName: \"kubernetes.io/projected/8e3486fc-aed7-4e82-8ad4-73b465917bbf-kube-api-access-c452f\") pod \"nova-metadata-0\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.513412 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.521648 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.524126 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.524198 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-svc\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.524253 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzg6f\" (UniqueName: \"kubernetes.io/projected/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-kube-api-access-nzg6f\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.524323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.524343 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-config\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.524366 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.525393 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.525993 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.526869 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-svc\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.527176 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.527623 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-config\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.555887 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.556991 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzg6f\" (UniqueName: \"kubernetes.io/projected/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-kube-api-access-nzg6f\") pod \"dnsmasq-dns-bccf8f775-f9t2n\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.596068 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w5n6b"] Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.626940 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:14 crc kubenswrapper[4758]: W1203 17:21:14.653287 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefe2b439_b61f_4a3e_accc_abc261fc216f.slice/crio-61343e32ffdfadf7e4688c103223c964b9535fca330e7369313754210773d42b WatchSource:0}: Error finding container 61343e32ffdfadf7e4688c103223c964b9535fca330e7369313754210773d42b: Status 404 returned error can't find the container with id 61343e32ffdfadf7e4688c103223c964b9535fca330e7369313754210773d42b Dec 03 17:21:14 crc kubenswrapper[4758]: I1203 17:21:14.777040 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.299337 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlkz4"] Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.302337 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.308523 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.309396 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.320258 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlkz4"] Dec 03 17:21:15 crc kubenswrapper[4758]: W1203 17:21:15.363218 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa876ee7_a6e5_4393_a387_a965eefb43dd.slice/crio-392a7438fd65103fc447d44872eab569d7ea44d5b2ec6c3dd62d45fd02946fac WatchSource:0}: Error finding container 392a7438fd65103fc447d44872eab569d7ea44d5b2ec6c3dd62d45fd02946fac: Status 404 returned error can't find the container with id 392a7438fd65103fc447d44872eab569d7ea44d5b2ec6c3dd62d45fd02946fac Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.376665 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.466584 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.466692 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6qwx\" (UniqueName: \"kubernetes.io/projected/96f73920-eaca-458d-8b54-375f8c0928a3-kube-api-access-c6qwx\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.466834 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.466867 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-scripts\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: W1203 17:21:15.533972 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bdf9ce3_1773_4441_b170_e0ac39fbb80c.slice/crio-a8c7b2bbadd7aa8b46e2be3385e2c903659abba0a6047ffdbcd7f7ba02b1863a WatchSource:0}: Error finding container a8c7b2bbadd7aa8b46e2be3385e2c903659abba0a6047ffdbcd7f7ba02b1863a: Status 404 returned error can't find the container with id a8c7b2bbadd7aa8b46e2be3385e2c903659abba0a6047ffdbcd7f7ba02b1863a Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.535113 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w5n6b" event={"ID":"efe2b439-b61f-4a3e-accc-abc261fc216f","Type":"ContainerStarted","Data":"2f0d4efaf8fb3419ad631f420a3c8d2ab75387d893dc5b9e0ec89e0193a7389c"} Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.535189 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w5n6b" event={"ID":"efe2b439-b61f-4a3e-accc-abc261fc216f","Type":"ContainerStarted","Data":"61343e32ffdfadf7e4688c103223c964b9535fca330e7369313754210773d42b"} Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.537526 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ca73448-4695-4227-8e0f-6a8a0a6e8e21","Type":"ContainerStarted","Data":"d68b1b5808f239be4350824e9cdd67c8cac075f6769105eb8faaa2bb1f031e3f"} Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.538324 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa876ee7-a6e5-4393-a387-a965eefb43dd","Type":"ContainerStarted","Data":"392a7438fd65103fc447d44872eab569d7ea44d5b2ec6c3dd62d45fd02946fac"} Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.539604 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.559010 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-f9t2n"] Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.568714 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-w5n6b" podStartSLOduration=2.568667101 podStartE2EDuration="2.568667101s" podCreationTimestamp="2025-12-03 17:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:15.565470365 +0000 UTC m=+1530.766847226" watchObservedRunningTime="2025-12-03 17:21:15.568667101 +0000 UTC m=+1530.770043962" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.568883 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.568999 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-scripts\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.569094 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.569154 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6qwx\" (UniqueName: \"kubernetes.io/projected/96f73920-eaca-458d-8b54-375f8c0928a3-kube-api-access-c6qwx\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.579764 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.580506 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-scripts\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.580873 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.593147 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6qwx\" (UniqueName: \"kubernetes.io/projected/96f73920-eaca-458d-8b54-375f8c0928a3-kube-api-access-c6qwx\") pod \"nova-cell1-conductor-db-sync-zlkz4\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.661763 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:15 crc kubenswrapper[4758]: I1203 17:21:15.663146 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.337513 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlkz4"] Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.573729 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bdf9ce3-1773-4441-b170-e0ac39fbb80c","Type":"ContainerStarted","Data":"a8c7b2bbadd7aa8b46e2be3385e2c903659abba0a6047ffdbcd7f7ba02b1863a"} Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.576895 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" event={"ID":"96f73920-eaca-458d-8b54-375f8c0928a3","Type":"ContainerStarted","Data":"4a57461a32cc6fcc73c9ba25d2691469d72368efbb675e6938f6d905e98a98d5"} Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.579838 4758 generic.go:334] "Generic (PLEG): container finished" podID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerID="725fb527456d51890a2e88aeb9e163111cb380a48930dbc8ce6dd902e06c3709" exitCode=0 Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.580150 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" event={"ID":"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e","Type":"ContainerDied","Data":"725fb527456d51890a2e88aeb9e163111cb380a48930dbc8ce6dd902e06c3709"} Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.582067 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" event={"ID":"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e","Type":"ContainerStarted","Data":"921ecb296b2d8fdd0a023ec1b13217dfadc266fa42b423b54fb974981544f50f"} Dec 03 17:21:16 crc kubenswrapper[4758]: I1203 17:21:16.594652 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e3486fc-aed7-4e82-8ad4-73b465917bbf","Type":"ContainerStarted","Data":"10fc6a6b623bdc11223843abcea2dda07f9623fb7c39e10311bc2153e7478c19"} Dec 03 17:21:17 crc kubenswrapper[4758]: I1203 17:21:17.638017 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" event={"ID":"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e","Type":"ContainerStarted","Data":"f8462cdf53aa4a2f1c9cf93174c1258d17adcfbc127e072e1de997ba0f6f330d"} Dec 03 17:21:17 crc kubenswrapper[4758]: I1203 17:21:17.642776 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:17 crc kubenswrapper[4758]: I1203 17:21:17.644031 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" event={"ID":"96f73920-eaca-458d-8b54-375f8c0928a3","Type":"ContainerStarted","Data":"28190b32c88efa920b93be379a781b27859fc7ef01a1ee77d97810a89f3906f3"} Dec 03 17:21:17 crc kubenswrapper[4758]: I1203 17:21:17.664110 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" podStartSLOduration=3.66408537 podStartE2EDuration="3.66408537s" podCreationTimestamp="2025-12-03 17:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:17.660423162 +0000 UTC m=+1532.861800023" watchObservedRunningTime="2025-12-03 17:21:17.66408537 +0000 UTC m=+1532.865462231" Dec 03 17:21:17 crc kubenswrapper[4758]: I1203 17:21:17.681230 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" podStartSLOduration=2.68120777 podStartE2EDuration="2.68120777s" podCreationTimestamp="2025-12-03 17:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:17.68007127 +0000 UTC m=+1532.881448121" watchObservedRunningTime="2025-12-03 17:21:17.68120777 +0000 UTC m=+1532.882584631" Dec 03 17:21:18 crc kubenswrapper[4758]: I1203 17:21:18.492838 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:18 crc kubenswrapper[4758]: I1203 17:21:18.535302 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:19 crc kubenswrapper[4758]: I1203 17:21:19.849279 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.727656 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ca73448-4695-4227-8e0f-6a8a0a6e8e21","Type":"ContainerStarted","Data":"ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e"} Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.728817 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0ca73448-4695-4227-8e0f-6a8a0a6e8e21" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e" gracePeriod=30 Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.733102 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa876ee7-a6e5-4393-a387-a965eefb43dd","Type":"ContainerStarted","Data":"c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1"} Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.756624 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e3486fc-aed7-4e82-8ad4-73b465917bbf","Type":"ContainerStarted","Data":"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f"} Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.762276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bdf9ce3-1773-4441-b170-e0ac39fbb80c","Type":"ContainerStarted","Data":"eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378"} Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.762279 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.466918861 podStartE2EDuration="7.762254488s" podCreationTimestamp="2025-12-03 17:21:13 +0000 UTC" firstStartedPulling="2025-12-03 17:21:14.900962254 +0000 UTC m=+1530.102339115" lastFinishedPulling="2025-12-03 17:21:20.196297881 +0000 UTC m=+1535.397674742" observedRunningTime="2025-12-03 17:21:20.757043237 +0000 UTC m=+1535.958420098" watchObservedRunningTime="2025-12-03 17:21:20.762254488 +0000 UTC m=+1535.963631349" Dec 03 17:21:20 crc kubenswrapper[4758]: I1203 17:21:20.802999 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.1894432 podStartE2EDuration="6.80296788s" podCreationTimestamp="2025-12-03 17:21:14 +0000 UTC" firstStartedPulling="2025-12-03 17:21:15.584944638 +0000 UTC m=+1530.786321499" lastFinishedPulling="2025-12-03 17:21:20.198469318 +0000 UTC m=+1535.399846179" observedRunningTime="2025-12-03 17:21:20.784094784 +0000 UTC m=+1535.985471645" watchObservedRunningTime="2025-12-03 17:21:20.80296788 +0000 UTC m=+1536.004344741" Dec 03 17:21:21 crc kubenswrapper[4758]: I1203 17:21:21.775422 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa876ee7-a6e5-4393-a387-a965eefb43dd","Type":"ContainerStarted","Data":"04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86"} Dec 03 17:21:21 crc kubenswrapper[4758]: I1203 17:21:21.777951 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e3486fc-aed7-4e82-8ad4-73b465917bbf","Type":"ContainerStarted","Data":"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496"} Dec 03 17:21:21 crc kubenswrapper[4758]: I1203 17:21:21.778222 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-log" containerID="cri-o://dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f" gracePeriod=30 Dec 03 17:21:21 crc kubenswrapper[4758]: I1203 17:21:21.778313 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-metadata" containerID="cri-o://77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496" gracePeriod=30 Dec 03 17:21:21 crc kubenswrapper[4758]: I1203 17:21:21.809265 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.980165258 podStartE2EDuration="8.809229673s" podCreationTimestamp="2025-12-03 17:21:13 +0000 UTC" firstStartedPulling="2025-12-03 17:21:15.366796364 +0000 UTC m=+1530.568173225" lastFinishedPulling="2025-12-03 17:21:20.195860779 +0000 UTC m=+1535.397237640" observedRunningTime="2025-12-03 17:21:21.80615686 +0000 UTC m=+1537.007533721" watchObservedRunningTime="2025-12-03 17:21:21.809229673 +0000 UTC m=+1537.010606534" Dec 03 17:21:21 crc kubenswrapper[4758]: I1203 17:21:21.845211 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.295707834 podStartE2EDuration="8.845177617s" podCreationTimestamp="2025-12-03 17:21:13 +0000 UTC" firstStartedPulling="2025-12-03 17:21:15.650703112 +0000 UTC m=+1530.852079973" lastFinishedPulling="2025-12-03 17:21:20.200172895 +0000 UTC m=+1535.401549756" observedRunningTime="2025-12-03 17:21:21.837319316 +0000 UTC m=+1537.038696187" watchObservedRunningTime="2025-12-03 17:21:21.845177617 +0000 UTC m=+1537.046554478" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.753752 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.791817 4758 generic.go:334] "Generic (PLEG): container finished" podID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerID="77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496" exitCode=0 Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.792260 4758 generic.go:334] "Generic (PLEG): container finished" podID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerID="dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f" exitCode=143 Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.793603 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.793664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e3486fc-aed7-4e82-8ad4-73b465917bbf","Type":"ContainerDied","Data":"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496"} Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.793775 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e3486fc-aed7-4e82-8ad4-73b465917bbf","Type":"ContainerDied","Data":"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f"} Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.793791 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e3486fc-aed7-4e82-8ad4-73b465917bbf","Type":"ContainerDied","Data":"10fc6a6b623bdc11223843abcea2dda07f9623fb7c39e10311bc2153e7478c19"} Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.793813 4758 scope.go:117] "RemoveContainer" containerID="77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.823891 4758 scope.go:117] "RemoveContainer" containerID="dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.853425 4758 scope.go:117] "RemoveContainer" containerID="77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496" Dec 03 17:21:22 crc kubenswrapper[4758]: E1203 17:21:22.856062 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496\": container with ID starting with 77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496 not found: ID does not exist" containerID="77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.856132 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496"} err="failed to get container status \"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496\": rpc error: code = NotFound desc = could not find container \"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496\": container with ID starting with 77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496 not found: ID does not exist" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.856171 4758 scope.go:117] "RemoveContainer" containerID="dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f" Dec 03 17:21:22 crc kubenswrapper[4758]: E1203 17:21:22.856655 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f\": container with ID starting with dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f not found: ID does not exist" containerID="dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.856729 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f"} err="failed to get container status \"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f\": rpc error: code = NotFound desc = could not find container \"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f\": container with ID starting with dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f not found: ID does not exist" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.856770 4758 scope.go:117] "RemoveContainer" containerID="77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.857074 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496"} err="failed to get container status \"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496\": rpc error: code = NotFound desc = could not find container \"77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496\": container with ID starting with 77b1d90ab4064858276f8d3572c98e7a7ba69f65e7bbd1880df8407a96cf6496 not found: ID does not exist" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.857100 4758 scope.go:117] "RemoveContainer" containerID="dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.857353 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f"} err="failed to get container status \"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f\": rpc error: code = NotFound desc = could not find container \"dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f\": container with ID starting with dc431a003860bd1726d709be0b87fe2fd3269426cc98219df4d6f912d4c8648f not found: ID does not exist" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.928174 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-config-data\") pod \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.928317 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-combined-ca-bundle\") pod \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.928501 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3486fc-aed7-4e82-8ad4-73b465917bbf-logs\") pod \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.928566 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c452f\" (UniqueName: \"kubernetes.io/projected/8e3486fc-aed7-4e82-8ad4-73b465917bbf-kube-api-access-c452f\") pod \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\" (UID: \"8e3486fc-aed7-4e82-8ad4-73b465917bbf\") " Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.929300 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e3486fc-aed7-4e82-8ad4-73b465917bbf-logs" (OuterVolumeSpecName: "logs") pod "8e3486fc-aed7-4e82-8ad4-73b465917bbf" (UID: "8e3486fc-aed7-4e82-8ad4-73b465917bbf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.930750 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3486fc-aed7-4e82-8ad4-73b465917bbf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.943499 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e3486fc-aed7-4e82-8ad4-73b465917bbf-kube-api-access-c452f" (OuterVolumeSpecName: "kube-api-access-c452f") pod "8e3486fc-aed7-4e82-8ad4-73b465917bbf" (UID: "8e3486fc-aed7-4e82-8ad4-73b465917bbf"). InnerVolumeSpecName "kube-api-access-c452f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.974451 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e3486fc-aed7-4e82-8ad4-73b465917bbf" (UID: "8e3486fc-aed7-4e82-8ad4-73b465917bbf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:22 crc kubenswrapper[4758]: I1203 17:21:22.980242 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-config-data" (OuterVolumeSpecName: "config-data") pod "8e3486fc-aed7-4e82-8ad4-73b465917bbf" (UID: "8e3486fc-aed7-4e82-8ad4-73b465917bbf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.032852 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c452f\" (UniqueName: \"kubernetes.io/projected/8e3486fc-aed7-4e82-8ad4-73b465917bbf-kube-api-access-c452f\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.032883 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.032903 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3486fc-aed7-4e82-8ad4-73b465917bbf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.140986 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.150804 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.180654 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:23 crc kubenswrapper[4758]: E1203 17:21:23.181206 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-log" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.181224 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-log" Dec 03 17:21:23 crc kubenswrapper[4758]: E1203 17:21:23.181254 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-metadata" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.181260 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-metadata" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.181458 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-metadata" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.181478 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" containerName="nova-metadata-log" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.182637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.186798 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.189079 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.203514 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.338878 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-config-data\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.338952 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.339730 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxk5w\" (UniqueName: \"kubernetes.io/projected/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-kube-api-access-pxk5w\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.339904 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.340218 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-logs\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.443206 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-config-data\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.443263 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.443314 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxk5w\" (UniqueName: \"kubernetes.io/projected/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-kube-api-access-pxk5w\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.443378 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.443418 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-logs\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.445741 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-logs\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.450056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.451264 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.451961 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-config-data\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.472094 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxk5w\" (UniqueName: \"kubernetes.io/projected/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-kube-api-access-pxk5w\") pod \"nova-metadata-0\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " pod="openstack/nova-metadata-0" Dec 03 17:21:23 crc kubenswrapper[4758]: I1203 17:21:23.507498 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.080239 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:24 crc kubenswrapper[4758]: W1203 17:21:24.097443 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c42c50_c98d_40cd_bf4d_f46492d3f16e.slice/crio-b42d710e02527ee833c7884f64723695dd07b624569e9c56ff4cc744943bd682 WatchSource:0}: Error finding container b42d710e02527ee833c7884f64723695dd07b624569e9c56ff4cc744943bd682: Status 404 returned error can't find the container with id b42d710e02527ee833c7884f64723695dd07b624569e9c56ff4cc744943bd682 Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.097471 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.514017 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.514099 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.556802 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.557254 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.617840 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.631232 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.753565 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7mn8j"] Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.754224 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerName="dnsmasq-dns" containerID="cri-o://24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e" gracePeriod=10 Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.842811 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4c42c50-c98d-40cd-bf4d-f46492d3f16e","Type":"ContainerStarted","Data":"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0"} Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.842902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4c42c50-c98d-40cd-bf4d-f46492d3f16e","Type":"ContainerStarted","Data":"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74"} Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.842918 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4c42c50-c98d-40cd-bf4d-f46492d3f16e","Type":"ContainerStarted","Data":"b42d710e02527ee833c7884f64723695dd07b624569e9c56ff4cc744943bd682"} Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.911041 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 17:21:24 crc kubenswrapper[4758]: I1203 17:21:24.968835 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.9688007870000002 podStartE2EDuration="1.968800787s" podCreationTimestamp="2025-12-03 17:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:24.875412132 +0000 UTC m=+1540.076789013" watchObservedRunningTime="2025-12-03 17:21:24.968800787 +0000 UTC m=+1540.170177648" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.142553 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e3486fc-aed7-4e82-8ad4-73b465917bbf" path="/var/lib/kubelet/pods/8e3486fc-aed7-4e82-8ad4-73b465917bbf/volumes" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.538272 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.597093 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.597114 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.713170 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-config\") pod \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.713405 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rtxs\" (UniqueName: \"kubernetes.io/projected/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-kube-api-access-6rtxs\") pod \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.713452 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-svc\") pod \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.713483 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-swift-storage-0\") pod \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.713526 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-sb\") pod \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.713578 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-nb\") pod \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\" (UID: \"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf\") " Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.753487 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-kube-api-access-6rtxs" (OuterVolumeSpecName: "kube-api-access-6rtxs") pod "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" (UID: "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf"). InnerVolumeSpecName "kube-api-access-6rtxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.812118 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" (UID: "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.816942 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rtxs\" (UniqueName: \"kubernetes.io/projected/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-kube-api-access-6rtxs\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.817002 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.825781 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" (UID: "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.829296 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" (UID: "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.844838 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-config" (OuterVolumeSpecName: "config") pod "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" (UID: "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.861975 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" (UID: "e60dc7d7-0f7a-4c68-9dc1-8637030b04bf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.865155 4758 generic.go:334] "Generic (PLEG): container finished" podID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerID="24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e" exitCode=0 Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.865304 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.865306 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" event={"ID":"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf","Type":"ContainerDied","Data":"24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e"} Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.865461 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7mn8j" event={"ID":"e60dc7d7-0f7a-4c68-9dc1-8637030b04bf","Type":"ContainerDied","Data":"9d0f04cb2200c4228efbb032998e4ed185c43285269a35f43d01c9bb3cf5a985"} Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.865515 4758 scope.go:117] "RemoveContainer" containerID="24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.928353 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.928399 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.928412 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.928427 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.971771 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7mn8j"] Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.974878 4758 scope.go:117] "RemoveContainer" containerID="96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16" Dec 03 17:21:25 crc kubenswrapper[4758]: I1203 17:21:25.991778 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7mn8j"] Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.020400 4758 scope.go:117] "RemoveContainer" containerID="24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e" Dec 03 17:21:26 crc kubenswrapper[4758]: E1203 17:21:26.027395 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e\": container with ID starting with 24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e not found: ID does not exist" containerID="24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e" Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.027448 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e"} err="failed to get container status \"24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e\": rpc error: code = NotFound desc = could not find container \"24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e\": container with ID starting with 24f86031e2272f86f4d744ed786f3b34ebc7b547bb54ba518f75ed8d2698930e not found: ID does not exist" Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.027482 4758 scope.go:117] "RemoveContainer" containerID="96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16" Dec 03 17:21:26 crc kubenswrapper[4758]: E1203 17:21:26.027954 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16\": container with ID starting with 96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16 not found: ID does not exist" containerID="96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16" Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.027988 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16"} err="failed to get container status \"96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16\": rpc error: code = NotFound desc = could not find container \"96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16\": container with ID starting with 96663219ac2d8d684788b19702f53ad49f944f7908bc564892cca595093aad16 not found: ID does not exist" Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.584727 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.585129 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" containerName="kube-state-metrics" containerID="cri-o://3e857659593ca56eeb3938a5a54b94b2417aab1032229a7c9c9ccbb10bc308db" gracePeriod=30 Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.919769 4758 generic.go:334] "Generic (PLEG): container finished" podID="efe2b439-b61f-4a3e-accc-abc261fc216f" containerID="2f0d4efaf8fb3419ad631f420a3c8d2ab75387d893dc5b9e0ec89e0193a7389c" exitCode=0 Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.919864 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w5n6b" event={"ID":"efe2b439-b61f-4a3e-accc-abc261fc216f","Type":"ContainerDied","Data":"2f0d4efaf8fb3419ad631f420a3c8d2ab75387d893dc5b9e0ec89e0193a7389c"} Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.930748 4758 generic.go:334] "Generic (PLEG): container finished" podID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" containerID="3e857659593ca56eeb3938a5a54b94b2417aab1032229a7c9c9ccbb10bc308db" exitCode=2 Dec 03 17:21:26 crc kubenswrapper[4758]: I1203 17:21:26.930806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4bbd2d2-37a4-46d1-b487-1ece965b4a11","Type":"ContainerDied","Data":"3e857659593ca56eeb3938a5a54b94b2417aab1032229a7c9c9ccbb10bc308db"} Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.138229 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" path="/var/lib/kubelet/pods/e60dc7d7-0f7a-4c68-9dc1-8637030b04bf/volumes" Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.169109 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.266095 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8njx\" (UniqueName: \"kubernetes.io/projected/d4bbd2d2-37a4-46d1-b487-1ece965b4a11-kube-api-access-f8njx\") pod \"d4bbd2d2-37a4-46d1-b487-1ece965b4a11\" (UID: \"d4bbd2d2-37a4-46d1-b487-1ece965b4a11\") " Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.286944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4bbd2d2-37a4-46d1-b487-1ece965b4a11-kube-api-access-f8njx" (OuterVolumeSpecName: "kube-api-access-f8njx") pod "d4bbd2d2-37a4-46d1-b487-1ece965b4a11" (UID: "d4bbd2d2-37a4-46d1-b487-1ece965b4a11"). InnerVolumeSpecName "kube-api-access-f8njx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.370138 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8njx\" (UniqueName: \"kubernetes.io/projected/d4bbd2d2-37a4-46d1-b487-1ece965b4a11-kube-api-access-f8njx\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.943414 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4bbd2d2-37a4-46d1-b487-1ece965b4a11","Type":"ContainerDied","Data":"c9066bdab2582fbfdb47955d74a5c388d9d9a627253f5e2117d96566265f0ae8"} Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.943474 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.943508 4758 scope.go:117] "RemoveContainer" containerID="3e857659593ca56eeb3938a5a54b94b2417aab1032229a7c9c9ccbb10bc308db" Dec 03 17:21:27 crc kubenswrapper[4758]: I1203 17:21:27.994053 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.012427 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.026488 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:21:28 crc kubenswrapper[4758]: E1203 17:21:28.027134 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerName="dnsmasq-dns" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.027172 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerName="dnsmasq-dns" Dec 03 17:21:28 crc kubenswrapper[4758]: E1203 17:21:28.027220 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" containerName="kube-state-metrics" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.027230 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" containerName="kube-state-metrics" Dec 03 17:21:28 crc kubenswrapper[4758]: E1203 17:21:28.027258 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerName="init" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.027285 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerName="init" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.027596 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e60dc7d7-0f7a-4c68-9dc1-8637030b04bf" containerName="dnsmasq-dns" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.027618 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" containerName="kube-state-metrics" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.028700 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.031736 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.032243 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.057256 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.192044 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v82zd\" (UniqueName: \"kubernetes.io/projected/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-api-access-v82zd\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.192302 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.192433 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.192475 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.295385 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v82zd\" (UniqueName: \"kubernetes.io/projected/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-api-access-v82zd\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.295458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.297047 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.297106 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.301532 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.313100 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.313438 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.314736 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v82zd\" (UniqueName: \"kubernetes.io/projected/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-api-access-v82zd\") pod \"kube-state-metrics-0\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.359741 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.463795 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.507990 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.509982 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.603156 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-combined-ca-bundle\") pod \"efe2b439-b61f-4a3e-accc-abc261fc216f\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.603251 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv4nt\" (UniqueName: \"kubernetes.io/projected/efe2b439-b61f-4a3e-accc-abc261fc216f-kube-api-access-xv4nt\") pod \"efe2b439-b61f-4a3e-accc-abc261fc216f\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.604650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-config-data\") pod \"efe2b439-b61f-4a3e-accc-abc261fc216f\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.604767 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-scripts\") pod \"efe2b439-b61f-4a3e-accc-abc261fc216f\" (UID: \"efe2b439-b61f-4a3e-accc-abc261fc216f\") " Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.627368 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-scripts" (OuterVolumeSpecName: "scripts") pod "efe2b439-b61f-4a3e-accc-abc261fc216f" (UID: "efe2b439-b61f-4a3e-accc-abc261fc216f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.627645 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe2b439-b61f-4a3e-accc-abc261fc216f-kube-api-access-xv4nt" (OuterVolumeSpecName: "kube-api-access-xv4nt") pod "efe2b439-b61f-4a3e-accc-abc261fc216f" (UID: "efe2b439-b61f-4a3e-accc-abc261fc216f"). InnerVolumeSpecName "kube-api-access-xv4nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.708216 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv4nt\" (UniqueName: \"kubernetes.io/projected/efe2b439-b61f-4a3e-accc-abc261fc216f-kube-api-access-xv4nt\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.708267 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.711881 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-config-data" (OuterVolumeSpecName: "config-data") pod "efe2b439-b61f-4a3e-accc-abc261fc216f" (UID: "efe2b439-b61f-4a3e-accc-abc261fc216f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.733935 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efe2b439-b61f-4a3e-accc-abc261fc216f" (UID: "efe2b439-b61f-4a3e-accc-abc261fc216f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.811344 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.811394 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2b439-b61f-4a3e-accc-abc261fc216f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.831556 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.956840 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82a1c0c2-3a27-43cc-8d41-d3c7508ab441","Type":"ContainerStarted","Data":"cf1ed8587f64f1208848cb64622fd7fe99899020ec341d6cf2a3359b753fc420"} Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.959399 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w5n6b" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.959450 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w5n6b" event={"ID":"efe2b439-b61f-4a3e-accc-abc261fc216f","Type":"ContainerDied","Data":"61343e32ffdfadf7e4688c103223c964b9535fca330e7369313754210773d42b"} Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.959509 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61343e32ffdfadf7e4688c103223c964b9535fca330e7369313754210773d42b" Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.963893 4758 generic.go:334] "Generic (PLEG): container finished" podID="96f73920-eaca-458d-8b54-375f8c0928a3" containerID="28190b32c88efa920b93be379a781b27859fc7ef01a1ee77d97810a89f3906f3" exitCode=0 Dec 03 17:21:28 crc kubenswrapper[4758]: I1203 17:21:28.963973 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" event={"ID":"96f73920-eaca-458d-8b54-375f8c0928a3","Type":"ContainerDied","Data":"28190b32c88efa920b93be379a781b27859fc7ef01a1ee77d97810a89f3906f3"} Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.097276 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.098283 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-log" containerID="cri-o://c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.098879 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-api" containerID="cri-o://04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.127061 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4bbd2d2-37a4-46d1-b487-1ece965b4a11" path="/var/lib/kubelet/pods/d4bbd2d2-37a4-46d1-b487-1ece965b4a11/volumes" Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.127719 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.127993 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" containerName="nova-scheduler-scheduler" containerID="cri-o://eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.131469 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.246774 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.247080 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-central-agent" containerID="cri-o://bba07a358e536f9d95a161647edd4d6a2ad3bfc0d86d90e9cc74f3fe0291df0b" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.247405 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="proxy-httpd" containerID="cri-o://c7c26dd0693f5732f58928023400c924fb34a8c2dcf2d6d240a6a91560f10d08" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.247502 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-notification-agent" containerID="cri-o://5d5d520d0a731b3f0bdb5b1c023550ac627838a6ed151ddec281961c090e9500" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: I1203 17:21:29.247602 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="sg-core" containerID="cri-o://7258d5fbd64e678990cf01affedc429932e83f78180f40d227f66e10a4188120" gracePeriod=30 Dec 03 17:21:29 crc kubenswrapper[4758]: E1203 17:21:29.564202 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:21:29 crc kubenswrapper[4758]: E1203 17:21:29.572747 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:21:29 crc kubenswrapper[4758]: E1203 17:21:29.582550 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:21:29 crc kubenswrapper[4758]: E1203 17:21:29.582653 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" containerName="nova-scheduler-scheduler" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.002930 4758 generic.go:334] "Generic (PLEG): container finished" podID="03107cd1-508b-404f-9728-43893717064b" containerID="7258d5fbd64e678990cf01affedc429932e83f78180f40d227f66e10a4188120" exitCode=2 Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.003378 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerDied","Data":"7258d5fbd64e678990cf01affedc429932e83f78180f40d227f66e10a4188120"} Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.010747 4758 generic.go:334] "Generic (PLEG): container finished" podID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerID="c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1" exitCode=143 Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.011275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa876ee7-a6e5-4393-a387-a965eefb43dd","Type":"ContainerDied","Data":"c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1"} Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.011628 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-log" containerID="cri-o://623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74" gracePeriod=30 Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.011935 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-metadata" containerID="cri-o://730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0" gracePeriod=30 Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.586112 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.715582 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.761625 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data\") pod \"96f73920-eaca-458d-8b54-375f8c0928a3\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.761784 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-combined-ca-bundle\") pod \"96f73920-eaca-458d-8b54-375f8c0928a3\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.761940 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6qwx\" (UniqueName: \"kubernetes.io/projected/96f73920-eaca-458d-8b54-375f8c0928a3-kube-api-access-c6qwx\") pod \"96f73920-eaca-458d-8b54-375f8c0928a3\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.761973 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-scripts\") pod \"96f73920-eaca-458d-8b54-375f8c0928a3\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.776917 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-scripts" (OuterVolumeSpecName: "scripts") pod "96f73920-eaca-458d-8b54-375f8c0928a3" (UID: "96f73920-eaca-458d-8b54-375f8c0928a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.778547 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96f73920-eaca-458d-8b54-375f8c0928a3-kube-api-access-c6qwx" (OuterVolumeSpecName: "kube-api-access-c6qwx") pod "96f73920-eaca-458d-8b54-375f8c0928a3" (UID: "96f73920-eaca-458d-8b54-375f8c0928a3"). InnerVolumeSpecName "kube-api-access-c6qwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: E1203 17:21:30.819137 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data podName:96f73920-eaca-458d-8b54-375f8c0928a3 nodeName:}" failed. No retries permitted until 2025-12-03 17:21:31.319091247 +0000 UTC m=+1546.520468108 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data") pod "96f73920-eaca-458d-8b54-375f8c0928a3" (UID: "96f73920-eaca-458d-8b54-375f8c0928a3") : error deleting /var/lib/kubelet/pods/96f73920-eaca-458d-8b54-375f8c0928a3/volume-subpaths: remove /var/lib/kubelet/pods/96f73920-eaca-458d-8b54-375f8c0928a3/volume-subpaths: no such file or directory Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.823397 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96f73920-eaca-458d-8b54-375f8c0928a3" (UID: "96f73920-eaca-458d-8b54-375f8c0928a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.864389 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-combined-ca-bundle\") pod \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.864565 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-logs\") pod \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.864648 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxk5w\" (UniqueName: \"kubernetes.io/projected/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-kube-api-access-pxk5w\") pod \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.864806 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-config-data\") pod \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.864864 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-nova-metadata-tls-certs\") pod \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\" (UID: \"d4c42c50-c98d-40cd-bf4d-f46492d3f16e\") " Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.865363 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6qwx\" (UniqueName: \"kubernetes.io/projected/96f73920-eaca-458d-8b54-375f8c0928a3-kube-api-access-c6qwx\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.865384 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.865394 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.866883 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-logs" (OuterVolumeSpecName: "logs") pod "d4c42c50-c98d-40cd-bf4d-f46492d3f16e" (UID: "d4c42c50-c98d-40cd-bf4d-f46492d3f16e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.870480 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-kube-api-access-pxk5w" (OuterVolumeSpecName: "kube-api-access-pxk5w") pod "d4c42c50-c98d-40cd-bf4d-f46492d3f16e" (UID: "d4c42c50-c98d-40cd-bf4d-f46492d3f16e"). InnerVolumeSpecName "kube-api-access-pxk5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.872655 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.912024 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-config-data" (OuterVolumeSpecName: "config-data") pod "d4c42c50-c98d-40cd-bf4d-f46492d3f16e" (UID: "d4c42c50-c98d-40cd-bf4d-f46492d3f16e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.913917 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4c42c50-c98d-40cd-bf4d-f46492d3f16e" (UID: "d4c42c50-c98d-40cd-bf4d-f46492d3f16e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.939910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d4c42c50-c98d-40cd-bf4d-f46492d3f16e" (UID: "d4c42c50-c98d-40cd-bf4d-f46492d3f16e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.967950 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.967999 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.968016 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxk5w\" (UniqueName: \"kubernetes.io/projected/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-kube-api-access-pxk5w\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.968031 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:30 crc kubenswrapper[4758]: I1203 17:21:30.968046 4758 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c42c50-c98d-40cd-bf4d-f46492d3f16e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.024235 4758 generic.go:334] "Generic (PLEG): container finished" podID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" exitCode=0 Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.024333 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.024355 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bdf9ce3-1773-4441-b170-e0ac39fbb80c","Type":"ContainerDied","Data":"eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.024812 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bdf9ce3-1773-4441-b170-e0ac39fbb80c","Type":"ContainerDied","Data":"a8c7b2bbadd7aa8b46e2be3385e2c903659abba0a6047ffdbcd7f7ba02b1863a"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.024852 4758 scope.go:117] "RemoveContainer" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.045018 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" event={"ID":"96f73920-eaca-458d-8b54-375f8c0928a3","Type":"ContainerDied","Data":"4a57461a32cc6fcc73c9ba25d2691469d72368efbb675e6938f6d905e98a98d5"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.045087 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a57461a32cc6fcc73c9ba25d2691469d72368efbb675e6938f6d905e98a98d5" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.045222 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlkz4" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.049388 4758 generic.go:334] "Generic (PLEG): container finished" podID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerID="730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0" exitCode=0 Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.049430 4758 generic.go:334] "Generic (PLEG): container finished" podID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerID="623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74" exitCode=143 Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.049486 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4c42c50-c98d-40cd-bf4d-f46492d3f16e","Type":"ContainerDied","Data":"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.049530 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4c42c50-c98d-40cd-bf4d-f46492d3f16e","Type":"ContainerDied","Data":"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.049547 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d4c42c50-c98d-40cd-bf4d-f46492d3f16e","Type":"ContainerDied","Data":"b42d710e02527ee833c7884f64723695dd07b624569e9c56ff4cc744943bd682"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.049665 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.070689 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-config-data\") pod \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.070794 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cgsx\" (UniqueName: \"kubernetes.io/projected/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-kube-api-access-8cgsx\") pod \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.071016 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-combined-ca-bundle\") pod \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\" (UID: \"8bdf9ce3-1773-4441-b170-e0ac39fbb80c\") " Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.071993 4758 generic.go:334] "Generic (PLEG): container finished" podID="03107cd1-508b-404f-9728-43893717064b" containerID="c7c26dd0693f5732f58928023400c924fb34a8c2dcf2d6d240a6a91560f10d08" exitCode=0 Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.072034 4758 generic.go:334] "Generic (PLEG): container finished" podID="03107cd1-508b-404f-9728-43893717064b" containerID="bba07a358e536f9d95a161647edd4d6a2ad3bfc0d86d90e9cc74f3fe0291df0b" exitCode=0 Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.072137 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerDied","Data":"c7c26dd0693f5732f58928023400c924fb34a8c2dcf2d6d240a6a91560f10d08"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.072186 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerDied","Data":"bba07a358e536f9d95a161647edd4d6a2ad3bfc0d86d90e9cc74f3fe0291df0b"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.077004 4758 scope.go:117] "RemoveContainer" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.078028 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378\": container with ID starting with eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378 not found: ID does not exist" containerID="eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.078142 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378"} err="failed to get container status \"eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378\": rpc error: code = NotFound desc = could not find container \"eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378\": container with ID starting with eab85ec9c8bef31dd65222251d9dd30273e332062b3889e8f84c48428b501378 not found: ID does not exist" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.078271 4758 scope.go:117] "RemoveContainer" containerID="730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.084622 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-kube-api-access-8cgsx" (OuterVolumeSpecName: "kube-api-access-8cgsx") pod "8bdf9ce3-1773-4441-b170-e0ac39fbb80c" (UID: "8bdf9ce3-1773-4441-b170-e0ac39fbb80c"). InnerVolumeSpecName "kube-api-access-8cgsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.084982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82a1c0c2-3a27-43cc-8d41-d3c7508ab441","Type":"ContainerStarted","Data":"f6091694aad9cdd8c04fa3979ee4bf04648cfc927b2b56cd01f71c7f65050343"} Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.087033 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.098937 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.100645 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-log" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.100699 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-log" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.100728 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-metadata" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.100738 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-metadata" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.100771 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe2b439-b61f-4a3e-accc-abc261fc216f" containerName="nova-manage" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.100783 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe2b439-b61f-4a3e-accc-abc261fc216f" containerName="nova-manage" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.100801 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" containerName="nova-scheduler-scheduler" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.100808 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" containerName="nova-scheduler-scheduler" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.100826 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f73920-eaca-458d-8b54-375f8c0928a3" containerName="nova-cell1-conductor-db-sync" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.100834 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f73920-eaca-458d-8b54-375f8c0928a3" containerName="nova-cell1-conductor-db-sync" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.101057 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-metadata" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.101072 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" containerName="nova-metadata-log" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.101088 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe2b439-b61f-4a3e-accc-abc261fc216f" containerName="nova-manage" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.101097 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" containerName="nova-scheduler-scheduler" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.101107 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="96f73920-eaca-458d-8b54-375f8c0928a3" containerName="nova-cell1-conductor-db-sync" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.102030 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.120821 4758 scope.go:117] "RemoveContainer" containerID="623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.174265 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-config-data" (OuterVolumeSpecName: "config-data") pod "8bdf9ce3-1773-4441-b170-e0ac39fbb80c" (UID: "8bdf9ce3-1773-4441-b170-e0ac39fbb80c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.178547 4758 scope.go:117] "RemoveContainer" containerID="730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.180545 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0\": container with ID starting with 730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0 not found: ID does not exist" containerID="730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.180578 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0"} err="failed to get container status \"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0\": rpc error: code = NotFound desc = could not find container \"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0\": container with ID starting with 730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0 not found: ID does not exist" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.180601 4758 scope.go:117] "RemoveContainer" containerID="623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.183635 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.183666 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.184939 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74\": container with ID starting with 623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74 not found: ID does not exist" containerID="623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.184972 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74"} err="failed to get container status \"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74\": rpc error: code = NotFound desc = could not find container \"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74\": container with ID starting with 623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74 not found: ID does not exist" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.184994 4758 scope.go:117] "RemoveContainer" containerID="730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.185983 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0"} err="failed to get container status \"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0\": rpc error: code = NotFound desc = could not find container \"730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0\": container with ID starting with 730a293e3205cb88ce497724e757e55a8188141477ee63e9e3657089700636c0 not found: ID does not exist" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.186031 4758 scope.go:117] "RemoveContainer" containerID="623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.186476 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74"} err="failed to get container status \"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74\": rpc error: code = NotFound desc = could not find container \"623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74\": container with ID starting with 623c0449dc1f7ba8db5525fe3034a5d6226c06cd172c382d36242b591eb0dd74 not found: ID does not exist" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.187644 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.187662 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cgsx\" (UniqueName: \"kubernetes.io/projected/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-kube-api-access-8cgsx\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.190835 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.202271 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bdf9ce3-1773-4441-b170-e0ac39fbb80c" (UID: "8bdf9ce3-1773-4441-b170-e0ac39fbb80c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.217448 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.220073 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.222563 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.223494 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.237148 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.272104 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.8875246519999997 podStartE2EDuration="4.272074982s" podCreationTimestamp="2025-12-03 17:21:27 +0000 UTC" firstStartedPulling="2025-12-03 17:21:28.818871511 +0000 UTC m=+1544.020248362" lastFinishedPulling="2025-12-03 17:21:29.203421831 +0000 UTC m=+1544.404798692" observedRunningTime="2025-12-03 17:21:31.197144471 +0000 UTC m=+1546.398521342" watchObservedRunningTime="2025-12-03 17:21:31.272074982 +0000 UTC m=+1546.473451843" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.289199 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.289276 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.289561 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7rhz\" (UniqueName: \"kubernetes.io/projected/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-kube-api-access-r7rhz\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.289693 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bdf9ce3-1773-4441-b170-e0ac39fbb80c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.391540 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data\") pod \"96f73920-eaca-458d-8b54-375f8c0928a3\" (UID: \"96f73920-eaca-458d-8b54-375f8c0928a3\") " Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392267 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-config-data\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392405 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxlvq\" (UniqueName: \"kubernetes.io/projected/7077de63-8c41-466f-b195-f1843f4e4d6b-kube-api-access-vxlvq\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392534 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7rhz\" (UniqueName: \"kubernetes.io/projected/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-kube-api-access-r7rhz\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392664 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392790 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392874 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.392965 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.393130 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7077de63-8c41-466f-b195-f1843f4e4d6b-logs\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.395128 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data" (OuterVolumeSpecName: "config-data") pod "96f73920-eaca-458d-8b54-375f8c0928a3" (UID: "96f73920-eaca-458d-8b54-375f8c0928a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.401276 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.401368 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.416636 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7rhz\" (UniqueName: \"kubernetes.io/projected/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-kube-api-access-r7rhz\") pod \"nova-cell1-conductor-0\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: E1203 17:21:31.424671 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c42c50_c98d_40cd_bf4d_f46492d3f16e.slice\": RecentStats: unable to find data in memory cache]" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.427788 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.445547 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.459126 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.461195 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.465071 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.473262 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.490099 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.495301 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.495447 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7077de63-8c41-466f-b195-f1843f4e4d6b-logs\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.495502 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-config-data\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.495528 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxlvq\" (UniqueName: \"kubernetes.io/projected/7077de63-8c41-466f-b195-f1843f4e4d6b-kube-api-access-vxlvq\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.495585 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.495651 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96f73920-eaca-458d-8b54-375f8c0928a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.499187 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7077de63-8c41-466f-b195-f1843f4e4d6b-logs\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.502794 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.503899 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-config-data\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.504250 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.521467 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxlvq\" (UniqueName: \"kubernetes.io/projected/7077de63-8c41-466f-b195-f1843f4e4d6b-kube-api-access-vxlvq\") pod \"nova-metadata-0\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.561333 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.606257 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-config-data\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.606444 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9v9t\" (UniqueName: \"kubernetes.io/projected/efb54870-24c2-4a71-962c-0a1f99725bd8-kube-api-access-q9v9t\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.606657 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.709582 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-config-data\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.709660 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9v9t\" (UniqueName: \"kubernetes.io/projected/efb54870-24c2-4a71-962c-0a1f99725bd8-kube-api-access-q9v9t\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.709826 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.725356 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.726006 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-config-data\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.732413 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9v9t\" (UniqueName: \"kubernetes.io/projected/efb54870-24c2-4a71-962c-0a1f99725bd8-kube-api-access-q9v9t\") pod \"nova-scheduler-0\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " pod="openstack/nova-scheduler-0" Dec 03 17:21:31 crc kubenswrapper[4758]: I1203 17:21:31.919584 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:21:32 crc kubenswrapper[4758]: I1203 17:21:32.119351 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:21:32 crc kubenswrapper[4758]: I1203 17:21:32.140310 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:21:32 crc kubenswrapper[4758]: I1203 17:21:32.464666 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:21:32 crc kubenswrapper[4758]: W1203 17:21:32.641894 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb54870_24c2_4a71_962c_0a1f99725bd8.slice/crio-ee6f0ddd39050e21860b25a1bcc21bdd1d6177f7b7b73208683a762a855e03b2 WatchSource:0}: Error finding container ee6f0ddd39050e21860b25a1bcc21bdd1d6177f7b7b73208683a762a855e03b2: Status 404 returned error can't find the container with id ee6f0ddd39050e21860b25a1bcc21bdd1d6177f7b7b73208683a762a855e03b2 Dec 03 17:21:32 crc kubenswrapper[4758]: I1203 17:21:32.960246 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.061507 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa876ee7-a6e5-4393-a387-a965eefb43dd-logs\") pod \"aa876ee7-a6e5-4393-a387-a965eefb43dd\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.061799 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-combined-ca-bundle\") pod \"aa876ee7-a6e5-4393-a387-a965eefb43dd\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.061895 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-config-data\") pod \"aa876ee7-a6e5-4393-a387-a965eefb43dd\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.061992 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq78b\" (UniqueName: \"kubernetes.io/projected/aa876ee7-a6e5-4393-a387-a965eefb43dd-kube-api-access-vq78b\") pod \"aa876ee7-a6e5-4393-a387-a965eefb43dd\" (UID: \"aa876ee7-a6e5-4393-a387-a965eefb43dd\") " Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.063081 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa876ee7-a6e5-4393-a387-a965eefb43dd-logs" (OuterVolumeSpecName: "logs") pod "aa876ee7-a6e5-4393-a387-a965eefb43dd" (UID: "aa876ee7-a6e5-4393-a387-a965eefb43dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.068928 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa876ee7-a6e5-4393-a387-a965eefb43dd-kube-api-access-vq78b" (OuterVolumeSpecName: "kube-api-access-vq78b") pod "aa876ee7-a6e5-4393-a387-a965eefb43dd" (UID: "aa876ee7-a6e5-4393-a387-a965eefb43dd"). InnerVolumeSpecName "kube-api-access-vq78b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.092877 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa876ee7-a6e5-4393-a387-a965eefb43dd" (UID: "aa876ee7-a6e5-4393-a387-a965eefb43dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.094616 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-config-data" (OuterVolumeSpecName: "config-data") pod "aa876ee7-a6e5-4393-a387-a965eefb43dd" (UID: "aa876ee7-a6e5-4393-a387-a965eefb43dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.134216 4758 generic.go:334] "Generic (PLEG): container finished" podID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerID="04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86" exitCode=0 Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.134795 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bdf9ce3-1773-4441-b170-e0ac39fbb80c" path="/var/lib/kubelet/pods/8bdf9ce3-1773-4441-b170-e0ac39fbb80c/volumes" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.135405 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c42c50-c98d-40cd-bf4d-f46492d3f16e" path="/var/lib/kubelet/pods/d4c42c50-c98d-40cd-bf4d-f46492d3f16e/volumes" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.135427 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136271 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136304 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"efb54870-24c2-4a71-962c-0a1f99725bd8","Type":"ContainerStarted","Data":"3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136323 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"efb54870-24c2-4a71-962c-0a1f99725bd8","Type":"ContainerStarted","Data":"ee6f0ddd39050e21860b25a1bcc21bdd1d6177f7b7b73208683a762a855e03b2"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136335 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b","Type":"ContainerStarted","Data":"c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b","Type":"ContainerStarted","Data":"94664d5ac08789bb46986d1262ce4f8a9b119515f5d6a7d6c618257f02119fdc"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136353 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7077de63-8c41-466f-b195-f1843f4e4d6b","Type":"ContainerStarted","Data":"aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7077de63-8c41-466f-b195-f1843f4e4d6b","Type":"ContainerStarted","Data":"b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136375 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7077de63-8c41-466f-b195-f1843f4e4d6b","Type":"ContainerStarted","Data":"1472d73faa01159d2940e9df6883aa12a46cff9daf529ac52e98dda073ef4f45"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136387 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa876ee7-a6e5-4393-a387-a965eefb43dd","Type":"ContainerDied","Data":"04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136402 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa876ee7-a6e5-4393-a387-a965eefb43dd","Type":"ContainerDied","Data":"392a7438fd65103fc447d44872eab569d7ea44d5b2ec6c3dd62d45fd02946fac"} Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.136443 4758 scope.go:117] "RemoveContainer" containerID="04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.164817 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa876ee7-a6e5-4393-a387-a965eefb43dd-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.164856 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.164870 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa876ee7-a6e5-4393-a387-a965eefb43dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.164884 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq78b\" (UniqueName: \"kubernetes.io/projected/aa876ee7-a6e5-4393-a387-a965eefb43dd-kube-api-access-vq78b\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.176034 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.175997523 podStartE2EDuration="2.175997523s" podCreationTimestamp="2025-12-03 17:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:33.172587841 +0000 UTC m=+1548.373964722" watchObservedRunningTime="2025-12-03 17:21:33.175997523 +0000 UTC m=+1548.377374384" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.177913 4758 scope.go:117] "RemoveContainer" containerID="c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.182906 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.182867247 podStartE2EDuration="2.182867247s" podCreationTimestamp="2025-12-03 17:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:33.147095537 +0000 UTC m=+1548.348472398" watchObservedRunningTime="2025-12-03 17:21:33.182867247 +0000 UTC m=+1548.384244108" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.199628 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.199592046 podStartE2EDuration="2.199592046s" podCreationTimestamp="2025-12-03 17:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:33.198174138 +0000 UTC m=+1548.399550999" watchObservedRunningTime="2025-12-03 17:21:33.199592046 +0000 UTC m=+1548.400968927" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.209982 4758 scope.go:117] "RemoveContainer" containerID="04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86" Dec 03 17:21:33 crc kubenswrapper[4758]: E1203 17:21:33.210505 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86\": container with ID starting with 04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86 not found: ID does not exist" containerID="04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.210555 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86"} err="failed to get container status \"04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86\": rpc error: code = NotFound desc = could not find container \"04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86\": container with ID starting with 04cd8d4573d5888cd3a5126861a5facd11db2cd19eec2bdf730e65f6d2507d86 not found: ID does not exist" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.210592 4758 scope.go:117] "RemoveContainer" containerID="c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1" Dec 03 17:21:33 crc kubenswrapper[4758]: E1203 17:21:33.211032 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1\": container with ID starting with c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1 not found: ID does not exist" containerID="c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.211112 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1"} err="failed to get container status \"c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1\": rpc error: code = NotFound desc = could not find container \"c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1\": container with ID starting with c18fc089e7c093222e09077875a483fa005ead5b7f12df9b5be6d740dc17d0f1 not found: ID does not exist" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.234793 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.274784 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.287486 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:33 crc kubenswrapper[4758]: E1203 17:21:33.288900 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-log" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.288929 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-log" Dec 03 17:21:33 crc kubenswrapper[4758]: E1203 17:21:33.288941 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-api" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.288950 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-api" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.289253 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-api" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.289293 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" containerName="nova-api-log" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.291063 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.306487 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.317147 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.380505 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.380591 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381cc840-d726-4e04-8549-9218f2dd30a0-logs\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.380644 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-config-data\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.380727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4xtr\" (UniqueName: \"kubernetes.io/projected/381cc840-d726-4e04-8549-9218f2dd30a0-kube-api-access-v4xtr\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.482316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.482398 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381cc840-d726-4e04-8549-9218f2dd30a0-logs\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.482481 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-config-data\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.483056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381cc840-d726-4e04-8549-9218f2dd30a0-logs\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.483802 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4xtr\" (UniqueName: \"kubernetes.io/projected/381cc840-d726-4e04-8549-9218f2dd30a0-kube-api-access-v4xtr\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.489239 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-config-data\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.495956 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.504529 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4xtr\" (UniqueName: \"kubernetes.io/projected/381cc840-d726-4e04-8549-9218f2dd30a0-kube-api-access-v4xtr\") pod \"nova-api-0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " pod="openstack/nova-api-0" Dec 03 17:21:33 crc kubenswrapper[4758]: I1203 17:21:33.643206 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:21:34 crc kubenswrapper[4758]: I1203 17:21:34.159509 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:35 crc kubenswrapper[4758]: I1203 17:21:35.129206 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa876ee7-a6e5-4393-a387-a965eefb43dd" path="/var/lib/kubelet/pods/aa876ee7-a6e5-4393-a387-a965eefb43dd/volumes" Dec 03 17:21:35 crc kubenswrapper[4758]: I1203 17:21:35.171833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381cc840-d726-4e04-8549-9218f2dd30a0","Type":"ContainerStarted","Data":"7429356e59a40130562703b88320ce49bcb235ccdfb6e0cc6a65ac09a2694e10"} Dec 03 17:21:35 crc kubenswrapper[4758]: I1203 17:21:35.171905 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381cc840-d726-4e04-8549-9218f2dd30a0","Type":"ContainerStarted","Data":"ca1e0244f903b8983e8975b79a8b3fcabc6737746219281259a5911c283e1f91"} Dec 03 17:21:35 crc kubenswrapper[4758]: I1203 17:21:35.171915 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381cc840-d726-4e04-8549-9218f2dd30a0","Type":"ContainerStarted","Data":"7ae859dc7e9c6246cdd1be229dbc5c2caca40f070c77d9f3ee149d9bb1126556"} Dec 03 17:21:35 crc kubenswrapper[4758]: I1203 17:21:35.195246 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.195211707 podStartE2EDuration="2.195211707s" podCreationTimestamp="2025-12-03 17:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:35.188521627 +0000 UTC m=+1550.389898508" watchObservedRunningTime="2025-12-03 17:21:35.195211707 +0000 UTC m=+1550.396588568" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.197078 4758 generic.go:334] "Generic (PLEG): container finished" podID="03107cd1-508b-404f-9728-43893717064b" containerID="5d5d520d0a731b3f0bdb5b1c023550ac627838a6ed151ddec281961c090e9500" exitCode=0 Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.197139 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerDied","Data":"5d5d520d0a731b3f0bdb5b1c023550ac627838a6ed151ddec281961c090e9500"} Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.563777 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.564281 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.658911 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809263 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-config-data\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809336 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-scripts\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809404 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-run-httpd\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809471 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-sg-core-conf-yaml\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809494 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-combined-ca-bundle\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-log-httpd\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809704 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2dzl\" (UniqueName: \"kubernetes.io/projected/03107cd1-508b-404f-9728-43893717064b-kube-api-access-g2dzl\") pod \"03107cd1-508b-404f-9728-43893717064b\" (UID: \"03107cd1-508b-404f-9728-43893717064b\") " Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.809982 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.810287 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.810327 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.824015 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-scripts" (OuterVolumeSpecName: "scripts") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.843286 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03107cd1-508b-404f-9728-43893717064b-kube-api-access-g2dzl" (OuterVolumeSpecName: "kube-api-access-g2dzl") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "kube-api-access-g2dzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.912135 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.916231 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03107cd1-508b-404f-9728-43893717064b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.916295 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2dzl\" (UniqueName: \"kubernetes.io/projected/03107cd1-508b-404f-9728-43893717064b-kube-api-access-g2dzl\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.916312 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.916323 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.920907 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.934316 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:36 crc kubenswrapper[4758]: I1203 17:21:36.984365 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-config-data" (OuterVolumeSpecName: "config-data") pod "03107cd1-508b-404f-9728-43893717064b" (UID: "03107cd1-508b-404f-9728-43893717064b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.019653 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.019722 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03107cd1-508b-404f-9728-43893717064b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.211915 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03107cd1-508b-404f-9728-43893717064b","Type":"ContainerDied","Data":"15876892b9da8a327b729f697291d12ccc619a91e6d8d565a0aa68c167ebe668"} Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.211987 4758 scope.go:117] "RemoveContainer" containerID="c7c26dd0693f5732f58928023400c924fb34a8c2dcf2d6d240a6a91560f10d08" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.212251 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.242002 4758 scope.go:117] "RemoveContainer" containerID="7258d5fbd64e678990cf01affedc429932e83f78180f40d227f66e10a4188120" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.248961 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.262987 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.278524 4758 scope.go:117] "RemoveContainer" containerID="5d5d520d0a731b3f0bdb5b1c023550ac627838a6ed151ddec281961c090e9500" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.288602 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:37 crc kubenswrapper[4758]: E1203 17:21:37.289347 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-notification-agent" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289376 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-notification-agent" Dec 03 17:21:37 crc kubenswrapper[4758]: E1203 17:21:37.289420 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="proxy-httpd" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289431 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="proxy-httpd" Dec 03 17:21:37 crc kubenswrapper[4758]: E1203 17:21:37.289451 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-central-agent" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289460 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-central-agent" Dec 03 17:21:37 crc kubenswrapper[4758]: E1203 17:21:37.289476 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="sg-core" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289483 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="sg-core" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289780 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-central-agent" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289801 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="sg-core" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289825 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="proxy-httpd" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.289837 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="03107cd1-508b-404f-9728-43893717064b" containerName="ceilometer-notification-agent" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.293765 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.304497 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.306536 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.306817 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.306964 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.311070 4758 scope.go:117] "RemoveContainer" containerID="bba07a358e536f9d95a161647edd4d6a2ad3bfc0d86d90e9cc74f3fe0291df0b" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.433811 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-scripts\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.433911 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-run-httpd\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.433974 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-log-httpd\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.434028 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-config-data\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.434138 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q22fw\" (UniqueName: \"kubernetes.io/projected/71e77daa-23bf-4df4-bcc9-f7b4bb072366-kube-api-access-q22fw\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.434171 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.434190 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.434374 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536310 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-log-httpd\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536395 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-config-data\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536479 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q22fw\" (UniqueName: \"kubernetes.io/projected/71e77daa-23bf-4df4-bcc9-f7b4bb072366-kube-api-access-q22fw\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536528 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536563 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536598 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536645 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-scripts\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.536744 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-run-httpd\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.537031 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-log-httpd\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.544946 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-run-httpd\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.546530 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.549308 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.550079 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-scripts\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.552705 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-config-data\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.558940 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q22fw\" (UniqueName: \"kubernetes.io/projected/71e77daa-23bf-4df4-bcc9-f7b4bb072366-kube-api-access-q22fw\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.559303 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " pod="openstack/ceilometer-0" Dec 03 17:21:37 crc kubenswrapper[4758]: I1203 17:21:37.638202 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:21:38 crc kubenswrapper[4758]: W1203 17:21:38.169279 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71e77daa_23bf_4df4_bcc9_f7b4bb072366.slice/crio-76276c173d44aeddabecf688bb9bdb82d0c9c247eef90d00e7537310f8bc9118 WatchSource:0}: Error finding container 76276c173d44aeddabecf688bb9bdb82d0c9c247eef90d00e7537310f8bc9118: Status 404 returned error can't find the container with id 76276c173d44aeddabecf688bb9bdb82d0c9c247eef90d00e7537310f8bc9118 Dec 03 17:21:38 crc kubenswrapper[4758]: I1203 17:21:38.170910 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:38 crc kubenswrapper[4758]: I1203 17:21:38.227350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerStarted","Data":"76276c173d44aeddabecf688bb9bdb82d0c9c247eef90d00e7537310f8bc9118"} Dec 03 17:21:38 crc kubenswrapper[4758]: I1203 17:21:38.369819 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 17:21:39 crc kubenswrapper[4758]: I1203 17:21:39.126418 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03107cd1-508b-404f-9728-43893717064b" path="/var/lib/kubelet/pods/03107cd1-508b-404f-9728-43893717064b/volumes" Dec 03 17:21:39 crc kubenswrapper[4758]: I1203 17:21:39.241838 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerStarted","Data":"d1d1f2ea885a87a1937f1c87ac261fc1f00fbee638788cff04570074d2c272fa"} Dec 03 17:21:40 crc kubenswrapper[4758]: I1203 17:21:40.258711 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerStarted","Data":"a3b0e7acbc1727ad6f9762d676ee3865d073829a02a625bf67ae229b4da00b74"} Dec 03 17:21:40 crc kubenswrapper[4758]: I1203 17:21:40.259134 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerStarted","Data":"10e7150de1b1e2cdb3b51882d60fcfeb3fb457d95fae7653a631c1fe5cb32c65"} Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.395973 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.396376 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.541992 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.563699 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.563778 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.920210 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 17:21:41 crc kubenswrapper[4758]: I1203 17:21:41.962308 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 17:21:42 crc kubenswrapper[4758]: I1203 17:21:42.286893 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerStarted","Data":"1fcb33c6b6383c85bf64ce3251e1a301a237347beb1c18c42a9dbcee78a92185"} Dec 03 17:21:42 crc kubenswrapper[4758]: I1203 17:21:42.286976 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:21:42 crc kubenswrapper[4758]: I1203 17:21:42.324778 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 17:21:42 crc kubenswrapper[4758]: I1203 17:21:42.373939 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.169142765 podStartE2EDuration="5.373915344s" podCreationTimestamp="2025-12-03 17:21:37 +0000 UTC" firstStartedPulling="2025-12-03 17:21:38.172501521 +0000 UTC m=+1553.373878382" lastFinishedPulling="2025-12-03 17:21:41.3772741 +0000 UTC m=+1556.578650961" observedRunningTime="2025-12-03 17:21:42.341850795 +0000 UTC m=+1557.543227666" watchObservedRunningTime="2025-12-03 17:21:42.373915344 +0000 UTC m=+1557.575292205" Dec 03 17:21:42 crc kubenswrapper[4758]: I1203 17:21:42.587979 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:21:42 crc kubenswrapper[4758]: I1203 17:21:42.588440 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:21:43 crc kubenswrapper[4758]: I1203 17:21:43.644155 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:21:43 crc kubenswrapper[4758]: I1203 17:21:43.644617 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:21:44 crc kubenswrapper[4758]: I1203 17:21:44.726054 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:21:44 crc kubenswrapper[4758]: I1203 17:21:44.726060 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.207900 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.392874 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-config-data\") pod \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.393118 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k75kz\" (UniqueName: \"kubernetes.io/projected/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-kube-api-access-k75kz\") pod \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.393255 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-combined-ca-bundle\") pod \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\" (UID: \"0ca73448-4695-4227-8e0f-6a8a0a6e8e21\") " Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.399389 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-kube-api-access-k75kz" (OuterVolumeSpecName: "kube-api-access-k75kz") pod "0ca73448-4695-4227-8e0f-6a8a0a6e8e21" (UID: "0ca73448-4695-4227-8e0f-6a8a0a6e8e21"). InnerVolumeSpecName "kube-api-access-k75kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.411934 4758 generic.go:334] "Generic (PLEG): container finished" podID="0ca73448-4695-4227-8e0f-6a8a0a6e8e21" containerID="ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e" exitCode=137 Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.411990 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ca73448-4695-4227-8e0f-6a8a0a6e8e21","Type":"ContainerDied","Data":"ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e"} Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.412022 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ca73448-4695-4227-8e0f-6a8a0a6e8e21","Type":"ContainerDied","Data":"d68b1b5808f239be4350824e9cdd67c8cac075f6769105eb8faaa2bb1f031e3f"} Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.412039 4758 scope.go:117] "RemoveContainer" containerID="ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.412169 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.426499 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ca73448-4695-4227-8e0f-6a8a0a6e8e21" (UID: "0ca73448-4695-4227-8e0f-6a8a0a6e8e21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.430803 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-config-data" (OuterVolumeSpecName: "config-data") pod "0ca73448-4695-4227-8e0f-6a8a0a6e8e21" (UID: "0ca73448-4695-4227-8e0f-6a8a0a6e8e21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.495614 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k75kz\" (UniqueName: \"kubernetes.io/projected/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-kube-api-access-k75kz\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.495657 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.495668 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca73448-4695-4227-8e0f-6a8a0a6e8e21-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.510288 4758 scope.go:117] "RemoveContainer" containerID="ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e" Dec 03 17:21:51 crc kubenswrapper[4758]: E1203 17:21:51.511073 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e\": container with ID starting with ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e not found: ID does not exist" containerID="ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.511147 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e"} err="failed to get container status \"ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e\": rpc error: code = NotFound desc = could not find container \"ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e\": container with ID starting with ed1039973c4e9d7b3f4fdfc543625e8eef47cba3cf3efd451b080104809e271e not found: ID does not exist" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.570488 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.571910 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.577431 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.756334 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.777779 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.793838 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:51 crc kubenswrapper[4758]: E1203 17:21:51.794505 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca73448-4695-4227-8e0f-6a8a0a6e8e21" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.794531 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca73448-4695-4227-8e0f-6a8a0a6e8e21" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.794817 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca73448-4695-4227-8e0f-6a8a0a6e8e21" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.795969 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.798902 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.799044 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.805709 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.809157 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.904578 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.905419 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.905623 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97dq7\" (UniqueName: \"kubernetes.io/projected/6df131e1-f13f-4943-95a0-18086525f883-kube-api-access-97dq7\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.905783 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:51 crc kubenswrapper[4758]: I1203 17:21:51.905923 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.007927 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.009068 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.009156 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97dq7\" (UniqueName: \"kubernetes.io/projected/6df131e1-f13f-4943-95a0-18086525f883-kube-api-access-97dq7\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.009184 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.009214 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.012879 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.014074 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.026326 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.026991 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.030031 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97dq7\" (UniqueName: \"kubernetes.io/projected/6df131e1-f13f-4943-95a0-18086525f883-kube-api-access-97dq7\") pod \"nova-cell1-novncproxy-0\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.117560 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.430595 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:21:52 crc kubenswrapper[4758]: I1203 17:21:52.654490 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.127989 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca73448-4695-4227-8e0f-6a8a0a6e8e21" path="/var/lib/kubelet/pods/0ca73448-4695-4227-8e0f-6a8a0a6e8e21/volumes" Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.436879 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6df131e1-f13f-4943-95a0-18086525f883","Type":"ContainerStarted","Data":"6bf75a7b95cf3a4d7d3c7984e0780c10a5e1cb76fbe0816ad9b844b0ed2b0584"} Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.436971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6df131e1-f13f-4943-95a0-18086525f883","Type":"ContainerStarted","Data":"5338ad0f74275da643c904a93cbd3a21618838542f50e1ca0d45757f40ac759e"} Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.474811 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.474784583 podStartE2EDuration="2.474784583s" podCreationTimestamp="2025-12-03 17:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:53.459829442 +0000 UTC m=+1568.661206303" watchObservedRunningTime="2025-12-03 17:21:53.474784583 +0000 UTC m=+1568.676161444" Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.649606 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.651822 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.658172 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:21:53 crc kubenswrapper[4758]: I1203 17:21:53.663463 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.457103 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.460666 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.670552 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-s582w"] Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.672951 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.683924 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-s582w"] Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.785703 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.785778 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l58rt\" (UniqueName: \"kubernetes.io/projected/091fea73-c280-4583-810c-1cfe5f1f0fe1-kube-api-access-l58rt\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.786146 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.786229 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-config\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.786404 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.786514 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.889477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.889582 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.889725 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.889751 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l58rt\" (UniqueName: \"kubernetes.io/projected/091fea73-c280-4583-810c-1cfe5f1f0fe1-kube-api-access-l58rt\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.890803 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.890839 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.891044 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.891068 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.891754 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.891815 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-config\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.892400 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-config\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:54 crc kubenswrapper[4758]: I1203 17:21:54.916884 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l58rt\" (UniqueName: \"kubernetes.io/projected/091fea73-c280-4583-810c-1cfe5f1f0fe1-kube-api-access-l58rt\") pod \"dnsmasq-dns-cd5cbd7b9-s582w\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:55 crc kubenswrapper[4758]: I1203 17:21:55.001468 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:55 crc kubenswrapper[4758]: I1203 17:21:55.530000 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-s582w"] Dec 03 17:21:56 crc kubenswrapper[4758]: I1203 17:21:56.480695 4758 generic.go:334] "Generic (PLEG): container finished" podID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerID="d2e1df737da67a9872c32625a37897f90c909e09e9f108e0525e5a4192296e68" exitCode=0 Dec 03 17:21:56 crc kubenswrapper[4758]: I1203 17:21:56.480826 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" event={"ID":"091fea73-c280-4583-810c-1cfe5f1f0fe1","Type":"ContainerDied","Data":"d2e1df737da67a9872c32625a37897f90c909e09e9f108e0525e5a4192296e68"} Dec 03 17:21:56 crc kubenswrapper[4758]: I1203 17:21:56.481222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" event={"ID":"091fea73-c280-4583-810c-1cfe5f1f0fe1","Type":"ContainerStarted","Data":"c023e3687670c040ecfddf0757e36f9bfa704e992f1bb359194cdaf3addb4a89"} Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.003924 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.004340 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-central-agent" containerID="cri-o://d1d1f2ea885a87a1937f1c87ac261fc1f00fbee638788cff04570074d2c272fa" gracePeriod=30 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.004440 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-notification-agent" containerID="cri-o://10e7150de1b1e2cdb3b51882d60fcfeb3fb457d95fae7653a631c1fe5cb32c65" gracePeriod=30 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.004472 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="proxy-httpd" containerID="cri-o://1fcb33c6b6383c85bf64ce3251e1a301a237347beb1c18c42a9dbcee78a92185" gracePeriod=30 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.004440 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="sg-core" containerID="cri-o://a3b0e7acbc1727ad6f9762d676ee3865d073829a02a625bf67ae229b4da00b74" gracePeriod=30 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.012638 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.198:3000/\": read tcp 10.217.0.2:49282->10.217.0.198:3000: read: connection reset by peer" Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.138865 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.493479 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" event={"ID":"091fea73-c280-4583-810c-1cfe5f1f0fe1","Type":"ContainerStarted","Data":"3c175f3c3abbd9f78b887a0d32c0fdf73322a6b77388cdd343be7a3ed9da0d73"} Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.493654 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.497252 4758 generic.go:334] "Generic (PLEG): container finished" podID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerID="1fcb33c6b6383c85bf64ce3251e1a301a237347beb1c18c42a9dbcee78a92185" exitCode=0 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.497293 4758 generic.go:334] "Generic (PLEG): container finished" podID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerID="a3b0e7acbc1727ad6f9762d676ee3865d073829a02a625bf67ae229b4da00b74" exitCode=2 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.497318 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerDied","Data":"1fcb33c6b6383c85bf64ce3251e1a301a237347beb1c18c42a9dbcee78a92185"} Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.497349 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerDied","Data":"a3b0e7acbc1727ad6f9762d676ee3865d073829a02a625bf67ae229b4da00b74"} Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.532385 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" podStartSLOduration=3.532359216 podStartE2EDuration="3.532359216s" podCreationTimestamp="2025-12-03 17:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:21:57.517080746 +0000 UTC m=+1572.718457607" watchObservedRunningTime="2025-12-03 17:21:57.532359216 +0000 UTC m=+1572.733736077" Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.974394 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.974654 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-log" containerID="cri-o://ca1e0244f903b8983e8975b79a8b3fcabc6737746219281259a5911c283e1f91" gracePeriod=30 Dec 03 17:21:57 crc kubenswrapper[4758]: I1203 17:21:57.974813 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-api" containerID="cri-o://7429356e59a40130562703b88320ce49bcb235ccdfb6e0cc6a65ac09a2694e10" gracePeriod=30 Dec 03 17:21:58 crc kubenswrapper[4758]: I1203 17:21:58.516050 4758 generic.go:334] "Generic (PLEG): container finished" podID="381cc840-d726-4e04-8549-9218f2dd30a0" containerID="ca1e0244f903b8983e8975b79a8b3fcabc6737746219281259a5911c283e1f91" exitCode=143 Dec 03 17:21:58 crc kubenswrapper[4758]: I1203 17:21:58.516655 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381cc840-d726-4e04-8549-9218f2dd30a0","Type":"ContainerDied","Data":"ca1e0244f903b8983e8975b79a8b3fcabc6737746219281259a5911c283e1f91"} Dec 03 17:21:58 crc kubenswrapper[4758]: I1203 17:21:58.521467 4758 generic.go:334] "Generic (PLEG): container finished" podID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerID="d1d1f2ea885a87a1937f1c87ac261fc1f00fbee638788cff04570074d2c272fa" exitCode=0 Dec 03 17:21:58 crc kubenswrapper[4758]: I1203 17:21:58.522477 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerDied","Data":"d1d1f2ea885a87a1937f1c87ac261fc1f00fbee638788cff04570074d2c272fa"} Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.558336 4758 generic.go:334] "Generic (PLEG): container finished" podID="381cc840-d726-4e04-8549-9218f2dd30a0" containerID="7429356e59a40130562703b88320ce49bcb235ccdfb6e0cc6a65ac09a2694e10" exitCode=0 Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.558416 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381cc840-d726-4e04-8549-9218f2dd30a0","Type":"ContainerDied","Data":"7429356e59a40130562703b88320ce49bcb235ccdfb6e0cc6a65ac09a2694e10"} Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.683284 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.782827 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381cc840-d726-4e04-8549-9218f2dd30a0-logs\") pod \"381cc840-d726-4e04-8549-9218f2dd30a0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.783046 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4xtr\" (UniqueName: \"kubernetes.io/projected/381cc840-d726-4e04-8549-9218f2dd30a0-kube-api-access-v4xtr\") pod \"381cc840-d726-4e04-8549-9218f2dd30a0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.783086 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-combined-ca-bundle\") pod \"381cc840-d726-4e04-8549-9218f2dd30a0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.783184 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-config-data\") pod \"381cc840-d726-4e04-8549-9218f2dd30a0\" (UID: \"381cc840-d726-4e04-8549-9218f2dd30a0\") " Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.785321 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/381cc840-d726-4e04-8549-9218f2dd30a0-logs" (OuterVolumeSpecName: "logs") pod "381cc840-d726-4e04-8549-9218f2dd30a0" (UID: "381cc840-d726-4e04-8549-9218f2dd30a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.790580 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/381cc840-d726-4e04-8549-9218f2dd30a0-kube-api-access-v4xtr" (OuterVolumeSpecName: "kube-api-access-v4xtr") pod "381cc840-d726-4e04-8549-9218f2dd30a0" (UID: "381cc840-d726-4e04-8549-9218f2dd30a0"). InnerVolumeSpecName "kube-api-access-v4xtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.824801 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "381cc840-d726-4e04-8549-9218f2dd30a0" (UID: "381cc840-d726-4e04-8549-9218f2dd30a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.829011 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-config-data" (OuterVolumeSpecName: "config-data") pod "381cc840-d726-4e04-8549-9218f2dd30a0" (UID: "381cc840-d726-4e04-8549-9218f2dd30a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.894365 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4xtr\" (UniqueName: \"kubernetes.io/projected/381cc840-d726-4e04-8549-9218f2dd30a0-kube-api-access-v4xtr\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.894415 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.894425 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/381cc840-d726-4e04-8549-9218f2dd30a0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:01 crc kubenswrapper[4758]: I1203 17:22:01.894434 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/381cc840-d726-4e04-8549-9218f2dd30a0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.123001 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.159835 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.578087 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"381cc840-d726-4e04-8549-9218f2dd30a0","Type":"ContainerDied","Data":"7ae859dc7e9c6246cdd1be229dbc5c2caca40f070c77d9f3ee149d9bb1126556"} Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.578447 4758 scope.go:117] "RemoveContainer" containerID="7429356e59a40130562703b88320ce49bcb235ccdfb6e0cc6a65ac09a2694e10" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.578587 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.589814 4758 generic.go:334] "Generic (PLEG): container finished" podID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerID="10e7150de1b1e2cdb3b51882d60fcfeb3fb457d95fae7653a631c1fe5cb32c65" exitCode=0 Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.589926 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerDied","Data":"10e7150de1b1e2cdb3b51882d60fcfeb3fb457d95fae7653a631c1fe5cb32c65"} Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.610635 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.729652 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.734585 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.741366 4758 scope.go:117] "RemoveContainer" containerID="ca1e0244f903b8983e8975b79a8b3fcabc6737746219281259a5911c283e1f91" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.748127 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.777753 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:02 crc kubenswrapper[4758]: E1203 17:22:02.778465 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="proxy-httpd" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778494 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="proxy-httpd" Dec 03 17:22:02 crc kubenswrapper[4758]: E1203 17:22:02.778527 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="sg-core" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778535 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="sg-core" Dec 03 17:22:02 crc kubenswrapper[4758]: E1203 17:22:02.778559 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-api" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778569 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-api" Dec 03 17:22:02 crc kubenswrapper[4758]: E1203 17:22:02.778580 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-notification-agent" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778587 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-notification-agent" Dec 03 17:22:02 crc kubenswrapper[4758]: E1203 17:22:02.778607 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-central-agent" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778615 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-central-agent" Dec 03 17:22:02 crc kubenswrapper[4758]: E1203 17:22:02.778634 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-log" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778641 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-log" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778930 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-notification-agent" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778954 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-log" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778967 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="proxy-httpd" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778981 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" containerName="nova-api-api" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.778994 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="sg-core" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.779049 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" containerName="ceilometer-central-agent" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.780459 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.786571 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.786930 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.787053 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836309 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-run-httpd\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836392 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-log-httpd\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-sg-core-conf-yaml\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836596 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-scripts\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836635 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-ceilometer-tls-certs\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836744 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-config-data\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836815 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-combined-ca-bundle\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.836880 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q22fw\" (UniqueName: \"kubernetes.io/projected/71e77daa-23bf-4df4-bcc9-f7b4bb072366-kube-api-access-q22fw\") pod \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\" (UID: \"71e77daa-23bf-4df4-bcc9-f7b4bb072366\") " Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.837587 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.837628 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.840413 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gsct\" (UniqueName: \"kubernetes.io/projected/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-kube-api-access-8gsct\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.840475 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-logs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.840777 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.840782 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.842426 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-config-data\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.842741 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.860641 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.882919 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.888089 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-scripts" (OuterVolumeSpecName: "scripts") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.985011 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e77daa-23bf-4df4-bcc9-f7b4bb072366-kube-api-access-q22fw" (OuterVolumeSpecName: "kube-api-access-q22fw") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "kube-api-access-q22fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.996638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.996709 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.996803 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gsct\" (UniqueName: \"kubernetes.io/projected/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-kube-api-access-8gsct\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.996835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-logs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.996911 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.996953 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-config-data\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.997052 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71e77daa-23bf-4df4-bcc9-f7b4bb072366-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.997106 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.997119 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q22fw\" (UniqueName: \"kubernetes.io/projected/71e77daa-23bf-4df4-bcc9-f7b4bb072366-kube-api-access-q22fw\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:02 crc kubenswrapper[4758]: I1203 17:22:02.999319 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-logs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.022239 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gsct\" (UniqueName: \"kubernetes.io/projected/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-kube-api-access-8gsct\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.023456 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-config-data\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.037228 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.039190 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.040427 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.044884 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-t442n"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.047490 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.052311 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.052587 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.056387 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.095492 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t442n"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.100062 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.117734 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.132070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.154993 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="381cc840-d726-4e04-8549-9218f2dd30a0" path="/var/lib/kubelet/pods/381cc840-d726-4e04-8549-9218f2dd30a0/volumes" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.202963 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.204703 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.204804 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lwgk\" (UniqueName: \"kubernetes.io/projected/a1f01fd7-322b-4576-86c8-83f8a06a7792-kube-api-access-9lwgk\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.204959 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-scripts\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.205056 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-config-data\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.205172 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.205193 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.259926 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-config-data" (OuterVolumeSpecName: "config-data") pod "71e77daa-23bf-4df4-bcc9-f7b4bb072366" (UID: "71e77daa-23bf-4df4-bcc9-f7b4bb072366"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.307085 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-config-data\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.307746 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.310935 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lwgk\" (UniqueName: \"kubernetes.io/projected/a1f01fd7-322b-4576-86c8-83f8a06a7792-kube-api-access-9lwgk\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.311347 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-scripts\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.311530 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e77daa-23bf-4df4-bcc9-f7b4bb072366-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.313287 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-config-data\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.317857 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.334100 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-scripts\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.349098 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lwgk\" (UniqueName: \"kubernetes.io/projected/a1f01fd7-322b-4576-86c8-83f8a06a7792-kube-api-access-9lwgk\") pod \"nova-cell1-cell-mapping-t442n\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.391693 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.625166 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71e77daa-23bf-4df4-bcc9-f7b4bb072366","Type":"ContainerDied","Data":"76276c173d44aeddabecf688bb9bdb82d0c9c247eef90d00e7537310f8bc9118"} Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.625242 4758 scope.go:117] "RemoveContainer" containerID="1fcb33c6b6383c85bf64ce3251e1a301a237347beb1c18c42a9dbcee78a92185" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.625410 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.674702 4758 scope.go:117] "RemoveContainer" containerID="a3b0e7acbc1727ad6f9762d676ee3865d073829a02a625bf67ae229b4da00b74" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.708915 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.737001 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.756412 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.760017 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.766417 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.766656 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.766812 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.774025 4758 scope.go:117] "RemoveContainer" containerID="10e7150de1b1e2cdb3b51882d60fcfeb3fb457d95fae7653a631c1fe5cb32c65" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.775695 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.802980 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.812485 4758 scope.go:117] "RemoveContainer" containerID="d1d1f2ea885a87a1937f1c87ac261fc1f00fbee638788cff04570074d2c272fa" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.932121 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-scripts\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.932177 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcgwg\" (UniqueName: \"kubernetes.io/projected/7bb44365-2dda-4769-813a-712f362fff6b-kube-api-access-lcgwg\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.932236 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-config-data\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.932440 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.932667 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.932792 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-run-httpd\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.933069 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-log-httpd\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:03 crc kubenswrapper[4758]: I1203 17:22:03.933147 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.035930 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-log-httpd\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.035992 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036072 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-scripts\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036103 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcgwg\" (UniqueName: \"kubernetes.io/projected/7bb44365-2dda-4769-813a-712f362fff6b-kube-api-access-lcgwg\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036134 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-config-data\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036173 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036216 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-run-httpd\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036491 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-log-httpd\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.036822 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-run-httpd\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.041285 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.041316 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.041471 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-config-data\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.043165 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-scripts\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.044277 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.065077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcgwg\" (UniqueName: \"kubernetes.io/projected/7bb44365-2dda-4769-813a-712f362fff6b-kube-api-access-lcgwg\") pod \"ceilometer-0\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.076707 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t442n"] Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.105007 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.637267 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.649133 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t442n" event={"ID":"a1f01fd7-322b-4576-86c8-83f8a06a7792","Type":"ContainerStarted","Data":"31c58b29312dd48a9c7ea2e97602a4641864e6411857864b3871e30db50be257"} Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.649218 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t442n" event={"ID":"a1f01fd7-322b-4576-86c8-83f8a06a7792","Type":"ContainerStarted","Data":"6c208b9f3a2a6a8f8404e53675a977a34e3c2a1724c0ac8f7741be30d637bd9d"} Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.676350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60b25a93-398b-4e93-a0fd-cb366c2bc3b2","Type":"ContainerStarted","Data":"20d40fb3d8e847fd076300ae39ef23864f01d587d60b52073c6509f653a1882c"} Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.676421 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60b25a93-398b-4e93-a0fd-cb366c2bc3b2","Type":"ContainerStarted","Data":"f3395389398c63f9417e3a4f3653aa710b138bf479b204f563ee182a23f6d4c6"} Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.676431 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60b25a93-398b-4e93-a0fd-cb366c2bc3b2","Type":"ContainerStarted","Data":"72e8e377ede390507999917e2a5eb8401ce5aa00bebd0d9e173825ab2ef75ba2"} Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.697212 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-t442n" podStartSLOduration=2.697174627 podStartE2EDuration="2.697174627s" podCreationTimestamp="2025-12-03 17:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:22:04.678969179 +0000 UTC m=+1579.880346050" watchObservedRunningTime="2025-12-03 17:22:04.697174627 +0000 UTC m=+1579.898551508" Dec 03 17:22:04 crc kubenswrapper[4758]: I1203 17:22:04.718928 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.718897598 podStartE2EDuration="2.718897598s" podCreationTimestamp="2025-12-03 17:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:22:04.707414651 +0000 UTC m=+1579.908791522" watchObservedRunningTime="2025-12-03 17:22:04.718897598 +0000 UTC m=+1579.920274459" Dec 03 17:22:05 crc kubenswrapper[4758]: I1203 17:22:05.003245 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:22:05 crc kubenswrapper[4758]: I1203 17:22:05.091066 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-f9t2n"] Dec 03 17:22:05 crc kubenswrapper[4758]: I1203 17:22:05.091409 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerName="dnsmasq-dns" containerID="cri-o://f8462cdf53aa4a2f1c9cf93174c1258d17adcfbc127e072e1de997ba0f6f330d" gracePeriod=10 Dec 03 17:22:05 crc kubenswrapper[4758]: I1203 17:22:05.172067 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71e77daa-23bf-4df4-bcc9-f7b4bb072366" path="/var/lib/kubelet/pods/71e77daa-23bf-4df4-bcc9-f7b4bb072366/volumes" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.718302 4758 generic.go:334] "Generic (PLEG): container finished" podID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerID="f8462cdf53aa4a2f1c9cf93174c1258d17adcfbc127e072e1de997ba0f6f330d" exitCode=0 Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.718707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" event={"ID":"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e","Type":"ContainerDied","Data":"f8462cdf53aa4a2f1c9cf93174c1258d17adcfbc127e072e1de997ba0f6f330d"} Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.718753 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" event={"ID":"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e","Type":"ContainerDied","Data":"921ecb296b2d8fdd0a023ec1b13217dfadc266fa42b423b54fb974981544f50f"} Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.718772 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="921ecb296b2d8fdd0a023ec1b13217dfadc266fa42b423b54fb974981544f50f" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.727185 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerStarted","Data":"e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388"} Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.727236 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerStarted","Data":"224ac65b99db035304592b13991230b1bbf407cc7a5745c110138c8c18b92e9e"} Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.740024 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.909001 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-swift-storage-0\") pod \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.909580 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-svc\") pod \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.909910 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-config\") pod \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.909957 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-nb\") pod \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.910136 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzg6f\" (UniqueName: \"kubernetes.io/projected/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-kube-api-access-nzg6f\") pod \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.910164 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-sb\") pod \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\" (UID: \"99d70ec4-5aee-49ce-b7dd-f4cc211ae77e\") " Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.924916 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-kube-api-access-nzg6f" (OuterVolumeSpecName: "kube-api-access-nzg6f") pod "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" (UID: "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e"). InnerVolumeSpecName "kube-api-access-nzg6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:05.995548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" (UID: "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.001593 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" (UID: "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.013155 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzg6f\" (UniqueName: \"kubernetes.io/projected/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-kube-api-access-nzg6f\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.013190 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.013200 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.016269 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" (UID: "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.016565 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-config" (OuterVolumeSpecName: "config") pod "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" (UID: "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.054420 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" (UID: "99d70ec4-5aee-49ce-b7dd-f4cc211ae77e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.117286 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.117318 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.117329 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.736819 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-f9t2n" Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.797042 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-f9t2n"] Dec 03 17:22:06 crc kubenswrapper[4758]: I1203 17:22:06.807433 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-f9t2n"] Dec 03 17:22:07 crc kubenswrapper[4758]: I1203 17:22:07.136883 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" path="/var/lib/kubelet/pods/99d70ec4-5aee-49ce-b7dd-f4cc211ae77e/volumes" Dec 03 17:22:07 crc kubenswrapper[4758]: I1203 17:22:07.753633 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerStarted","Data":"a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad"} Dec 03 17:22:07 crc kubenswrapper[4758]: I1203 17:22:07.755459 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerStarted","Data":"d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c"} Dec 03 17:22:09 crc kubenswrapper[4758]: I1203 17:22:09.781921 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerStarted","Data":"922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a"} Dec 03 17:22:09 crc kubenswrapper[4758]: I1203 17:22:09.782671 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:22:09 crc kubenswrapper[4758]: I1203 17:22:09.815417 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.617670091 podStartE2EDuration="6.815389578s" podCreationTimestamp="2025-12-03 17:22:03 +0000 UTC" firstStartedPulling="2025-12-03 17:22:04.658326566 +0000 UTC m=+1579.859703427" lastFinishedPulling="2025-12-03 17:22:08.856046053 +0000 UTC m=+1584.057422914" observedRunningTime="2025-12-03 17:22:09.807548399 +0000 UTC m=+1585.008925260" watchObservedRunningTime="2025-12-03 17:22:09.815389578 +0000 UTC m=+1585.016766439" Dec 03 17:22:09 crc kubenswrapper[4758]: I1203 17:22:09.962470 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 03 17:22:11 crc kubenswrapper[4758]: I1203 17:22:11.420491 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:22:11 crc kubenswrapper[4758]: I1203 17:22:11.420926 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:22:11 crc kubenswrapper[4758]: I1203 17:22:11.815529 4758 generic.go:334] "Generic (PLEG): container finished" podID="a1f01fd7-322b-4576-86c8-83f8a06a7792" containerID="31c58b29312dd48a9c7ea2e97602a4641864e6411857864b3871e30db50be257" exitCode=0 Dec 03 17:22:11 crc kubenswrapper[4758]: I1203 17:22:11.815592 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t442n" event={"ID":"a1f01fd7-322b-4576-86c8-83f8a06a7792","Type":"ContainerDied","Data":"31c58b29312dd48a9c7ea2e97602a4641864e6411857864b3871e30db50be257"} Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.135026 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.136743 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.215040 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.292369 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-config-data\") pod \"a1f01fd7-322b-4576-86c8-83f8a06a7792\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.292669 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-combined-ca-bundle\") pod \"a1f01fd7-322b-4576-86c8-83f8a06a7792\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.292733 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-scripts\") pod \"a1f01fd7-322b-4576-86c8-83f8a06a7792\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.292804 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lwgk\" (UniqueName: \"kubernetes.io/projected/a1f01fd7-322b-4576-86c8-83f8a06a7792-kube-api-access-9lwgk\") pod \"a1f01fd7-322b-4576-86c8-83f8a06a7792\" (UID: \"a1f01fd7-322b-4576-86c8-83f8a06a7792\") " Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.300409 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1f01fd7-322b-4576-86c8-83f8a06a7792-kube-api-access-9lwgk" (OuterVolumeSpecName: "kube-api-access-9lwgk") pod "a1f01fd7-322b-4576-86c8-83f8a06a7792" (UID: "a1f01fd7-322b-4576-86c8-83f8a06a7792"). InnerVolumeSpecName "kube-api-access-9lwgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.302987 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-scripts" (OuterVolumeSpecName: "scripts") pod "a1f01fd7-322b-4576-86c8-83f8a06a7792" (UID: "a1f01fd7-322b-4576-86c8-83f8a06a7792"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.334565 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-config-data" (OuterVolumeSpecName: "config-data") pod "a1f01fd7-322b-4576-86c8-83f8a06a7792" (UID: "a1f01fd7-322b-4576-86c8-83f8a06a7792"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.347233 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1f01fd7-322b-4576-86c8-83f8a06a7792" (UID: "a1f01fd7-322b-4576-86c8-83f8a06a7792"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.396247 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.396298 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.396312 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lwgk\" (UniqueName: \"kubernetes.io/projected/a1f01fd7-322b-4576-86c8-83f8a06a7792-kube-api-access-9lwgk\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.396326 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f01fd7-322b-4576-86c8-83f8a06a7792-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.867003 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t442n" event={"ID":"a1f01fd7-322b-4576-86c8-83f8a06a7792","Type":"ContainerDied","Data":"6c208b9f3a2a6a8f8404e53675a977a34e3c2a1724c0ac8f7741be30d637bd9d"} Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.867067 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t442n" Dec 03 17:22:13 crc kubenswrapper[4758]: I1203 17:22:13.867109 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c208b9f3a2a6a8f8404e53675a977a34e3c2a1724c0ac8f7741be30d637bd9d" Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.084060 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.084503 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="efb54870-24c2-4a71-962c-0a1f99725bd8" containerName="nova-scheduler-scheduler" containerID="cri-o://3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb" gracePeriod=30 Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.103906 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.119220 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.119638 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-log" containerID="cri-o://b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450" gracePeriod=30 Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.119753 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-metadata" containerID="cri-o://aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce" gracePeriod=30 Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.131960 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.131987 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.882848 4758 generic.go:334] "Generic (PLEG): container finished" podID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerID="b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450" exitCode=143 Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.882908 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7077de63-8c41-466f-b195-f1843f4e4d6b","Type":"ContainerDied","Data":"b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450"} Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.883401 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-log" containerID="cri-o://f3395389398c63f9417e3a4f3653aa710b138bf479b204f563ee182a23f6d4c6" gracePeriod=30 Dec 03 17:22:14 crc kubenswrapper[4758]: I1203 17:22:14.883551 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-api" containerID="cri-o://20d40fb3d8e847fd076300ae39ef23864f01d587d60b52073c6509f653a1882c" gracePeriod=30 Dec 03 17:22:15 crc kubenswrapper[4758]: I1203 17:22:15.898694 4758 generic.go:334] "Generic (PLEG): container finished" podID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerID="f3395389398c63f9417e3a4f3653aa710b138bf479b204f563ee182a23f6d4c6" exitCode=143 Dec 03 17:22:15 crc kubenswrapper[4758]: I1203 17:22:15.898773 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60b25a93-398b-4e93-a0fd-cb366c2bc3b2","Type":"ContainerDied","Data":"f3395389398c63f9417e3a4f3653aa710b138bf479b204f563ee182a23f6d4c6"} Dec 03 17:22:16 crc kubenswrapper[4758]: I1203 17:22:16.919399 4758 generic.go:334] "Generic (PLEG): container finished" podID="efb54870-24c2-4a71-962c-0a1f99725bd8" containerID="3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb" exitCode=0 Dec 03 17:22:16 crc kubenswrapper[4758]: I1203 17:22:16.919517 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"efb54870-24c2-4a71-962c-0a1f99725bd8","Type":"ContainerDied","Data":"3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb"} Dec 03 17:22:16 crc kubenswrapper[4758]: E1203 17:22:16.920540 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb is running failed: container process not found" containerID="3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:22:16 crc kubenswrapper[4758]: E1203 17:22:16.920977 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb is running failed: container process not found" containerID="3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:22:16 crc kubenswrapper[4758]: E1203 17:22:16.921383 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb is running failed: container process not found" containerID="3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:22:16 crc kubenswrapper[4758]: E1203 17:22:16.921431 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="efb54870-24c2-4a71-962c-0a1f99725bd8" containerName="nova-scheduler-scheduler" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.196074 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.308480 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-combined-ca-bundle\") pod \"efb54870-24c2-4a71-962c-0a1f99725bd8\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.308617 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-config-data\") pod \"efb54870-24c2-4a71-962c-0a1f99725bd8\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.308894 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9v9t\" (UniqueName: \"kubernetes.io/projected/efb54870-24c2-4a71-962c-0a1f99725bd8-kube-api-access-q9v9t\") pod \"efb54870-24c2-4a71-962c-0a1f99725bd8\" (UID: \"efb54870-24c2-4a71-962c-0a1f99725bd8\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.316218 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb54870-24c2-4a71-962c-0a1f99725bd8-kube-api-access-q9v9t" (OuterVolumeSpecName: "kube-api-access-q9v9t") pod "efb54870-24c2-4a71-962c-0a1f99725bd8" (UID: "efb54870-24c2-4a71-962c-0a1f99725bd8"). InnerVolumeSpecName "kube-api-access-q9v9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.342279 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:33756->10.217.0.195:8775: read: connection reset by peer" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.342725 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:33742->10.217.0.195:8775: read: connection reset by peer" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.347087 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-config-data" (OuterVolumeSpecName: "config-data") pod "efb54870-24c2-4a71-962c-0a1f99725bd8" (UID: "efb54870-24c2-4a71-962c-0a1f99725bd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.348368 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efb54870-24c2-4a71-962c-0a1f99725bd8" (UID: "efb54870-24c2-4a71-962c-0a1f99725bd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.411962 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.412017 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efb54870-24c2-4a71-962c-0a1f99725bd8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.412031 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9v9t\" (UniqueName: \"kubernetes.io/projected/efb54870-24c2-4a71-962c-0a1f99725bd8-kube-api-access-q9v9t\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.869764 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.921733 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-nova-metadata-tls-certs\") pod \"7077de63-8c41-466f-b195-f1843f4e4d6b\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.921786 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-combined-ca-bundle\") pod \"7077de63-8c41-466f-b195-f1843f4e4d6b\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.921845 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxlvq\" (UniqueName: \"kubernetes.io/projected/7077de63-8c41-466f-b195-f1843f4e4d6b-kube-api-access-vxlvq\") pod \"7077de63-8c41-466f-b195-f1843f4e4d6b\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.921889 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-config-data\") pod \"7077de63-8c41-466f-b195-f1843f4e4d6b\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.922079 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7077de63-8c41-466f-b195-f1843f4e4d6b-logs\") pod \"7077de63-8c41-466f-b195-f1843f4e4d6b\" (UID: \"7077de63-8c41-466f-b195-f1843f4e4d6b\") " Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.922956 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7077de63-8c41-466f-b195-f1843f4e4d6b-logs" (OuterVolumeSpecName: "logs") pod "7077de63-8c41-466f-b195-f1843f4e4d6b" (UID: "7077de63-8c41-466f-b195-f1843f4e4d6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:22:17 crc kubenswrapper[4758]: I1203 17:22:17.988952 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7077de63-8c41-466f-b195-f1843f4e4d6b-kube-api-access-vxlvq" (OuterVolumeSpecName: "kube-api-access-vxlvq") pod "7077de63-8c41-466f-b195-f1843f4e4d6b" (UID: "7077de63-8c41-466f-b195-f1843f4e4d6b"). InnerVolumeSpecName "kube-api-access-vxlvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.004055 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"efb54870-24c2-4a71-962c-0a1f99725bd8","Type":"ContainerDied","Data":"ee6f0ddd39050e21860b25a1bcc21bdd1d6177f7b7b73208683a762a855e03b2"} Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.004146 4758 scope.go:117] "RemoveContainer" containerID="3ce7317f70a2cc354d8cc7a796ffc3ba630ae007944d9c66a47c54d8ee6115eb" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.004376 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.025401 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7077de63-8c41-466f-b195-f1843f4e4d6b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.025438 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxlvq\" (UniqueName: \"kubernetes.io/projected/7077de63-8c41-466f-b195-f1843f4e4d6b-kube-api-access-vxlvq\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.070139 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7077de63-8c41-466f-b195-f1843f4e4d6b" (UID: "7077de63-8c41-466f-b195-f1843f4e4d6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.072083 4758 generic.go:334] "Generic (PLEG): container finished" podID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerID="aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce" exitCode=0 Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.072150 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7077de63-8c41-466f-b195-f1843f4e4d6b","Type":"ContainerDied","Data":"aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce"} Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.072189 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7077de63-8c41-466f-b195-f1843f4e4d6b","Type":"ContainerDied","Data":"1472d73faa01159d2940e9df6883aa12a46cff9daf529ac52e98dda073ef4f45"} Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.072293 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.076199 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-config-data" (OuterVolumeSpecName: "config-data") pod "7077de63-8c41-466f-b195-f1843f4e4d6b" (UID: "7077de63-8c41-466f-b195-f1843f4e4d6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.128927 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.128971 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.148001 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7077de63-8c41-466f-b195-f1843f4e4d6b" (UID: "7077de63-8c41-466f-b195-f1843f4e4d6b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.153935 4758 scope.go:117] "RemoveContainer" containerID="aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.187691 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.226948 4758 scope.go:117] "RemoveContainer" containerID="b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.233611 4758 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7077de63-8c41-466f-b195-f1843f4e4d6b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.234768 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.270436 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.271124 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-log" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271146 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-log" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.271186 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerName="init" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271194 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerName="init" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.271207 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb54870-24c2-4a71-962c-0a1f99725bd8" containerName="nova-scheduler-scheduler" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271216 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb54870-24c2-4a71-962c-0a1f99725bd8" containerName="nova-scheduler-scheduler" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.271227 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-metadata" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271234 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-metadata" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.271250 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1f01fd7-322b-4576-86c8-83f8a06a7792" containerName="nova-manage" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271257 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1f01fd7-322b-4576-86c8-83f8a06a7792" containerName="nova-manage" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.271272 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerName="dnsmasq-dns" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271281 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerName="dnsmasq-dns" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271592 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1f01fd7-322b-4576-86c8-83f8a06a7792" containerName="nova-manage" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271607 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-log" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271620 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" containerName="nova-metadata-metadata" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271633 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb54870-24c2-4a71-962c-0a1f99725bd8" containerName="nova-scheduler-scheduler" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.271650 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d70ec4-5aee-49ce-b7dd-f4cc211ae77e" containerName="dnsmasq-dns" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.272620 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.286044 4758 scope.go:117] "RemoveContainer" containerID="aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.286142 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.287012 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce\": container with ID starting with aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce not found: ID does not exist" containerID="aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.287056 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce"} err="failed to get container status \"aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce\": rpc error: code = NotFound desc = could not find container \"aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce\": container with ID starting with aadf3ae9f0d95e0d65b3cd78847282cd599b713f6558757a5661fdd94f8dc6ce not found: ID does not exist" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.287082 4758 scope.go:117] "RemoveContainer" containerID="b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450" Dec 03 17:22:18 crc kubenswrapper[4758]: E1203 17:22:18.287716 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450\": container with ID starting with b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450 not found: ID does not exist" containerID="b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.287735 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450"} err="failed to get container status \"b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450\": rpc error: code = NotFound desc = could not find container \"b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450\": container with ID starting with b5ea60d7ccab123c59bccfd8ecfc3097c149ab05d093027a0c9197c3bb34e450 not found: ID does not exist" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.295439 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.335901 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.336296 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p69sh\" (UniqueName: \"kubernetes.io/projected/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-kube-api-access-p69sh\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.336452 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-config-data\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.415710 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.433496 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.439249 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p69sh\" (UniqueName: \"kubernetes.io/projected/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-kube-api-access-p69sh\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.439336 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-config-data\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.439407 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.444458 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-config-data\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.445737 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.447096 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.449102 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.452066 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.452084 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.458444 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.464645 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p69sh\" (UniqueName: \"kubernetes.io/projected/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-kube-api-access-p69sh\") pod \"nova-scheduler-0\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.542487 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtltd\" (UniqueName: \"kubernetes.io/projected/84469de1-c706-4442-8610-31188889dad1-kube-api-access-xtltd\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.542714 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.542943 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.543076 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84469de1-c706-4442-8610-31188889dad1-logs\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.543150 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-config-data\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.620788 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.644316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.644383 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84469de1-c706-4442-8610-31188889dad1-logs\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.644413 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-config-data\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.644458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtltd\" (UniqueName: \"kubernetes.io/projected/84469de1-c706-4442-8610-31188889dad1-kube-api-access-xtltd\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.644525 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.646332 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84469de1-c706-4442-8610-31188889dad1-logs\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.649737 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.650247 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-config-data\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.651082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.664008 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtltd\" (UniqueName: \"kubernetes.io/projected/84469de1-c706-4442-8610-31188889dad1-kube-api-access-xtltd\") pod \"nova-metadata-0\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " pod="openstack/nova-metadata-0" Dec 03 17:22:18 crc kubenswrapper[4758]: I1203 17:22:18.780401 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:22:19 crc kubenswrapper[4758]: I1203 17:22:19.130626 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7077de63-8c41-466f-b195-f1843f4e4d6b" path="/var/lib/kubelet/pods/7077de63-8c41-466f-b195-f1843f4e4d6b/volumes" Dec 03 17:22:19 crc kubenswrapper[4758]: I1203 17:22:19.132130 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb54870-24c2-4a71-962c-0a1f99725bd8" path="/var/lib/kubelet/pods/efb54870-24c2-4a71-962c-0a1f99725bd8/volumes" Dec 03 17:22:19 crc kubenswrapper[4758]: I1203 17:22:19.145094 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:22:19 crc kubenswrapper[4758]: I1203 17:22:19.295718 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.104623 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04b3f0d2-7b01-48a6-b2be-9e205027d1a2","Type":"ContainerStarted","Data":"0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41"} Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.105112 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04b3f0d2-7b01-48a6-b2be-9e205027d1a2","Type":"ContainerStarted","Data":"ed8709865c5ffef4e83ed1e2430a6460910b2286370ca8d329cc488dfaf14eff"} Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.107719 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84469de1-c706-4442-8610-31188889dad1","Type":"ContainerStarted","Data":"08867454fb2022e6047c48092dc940ce2f60d0e3bb28b9647d08e3cc414e048e"} Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.107809 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84469de1-c706-4442-8610-31188889dad1","Type":"ContainerStarted","Data":"39dfeec538bbe7eb43d14ac4885d21560e65527c9acf6664be573a26763ad32b"} Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.107820 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84469de1-c706-4442-8610-31188889dad1","Type":"ContainerStarted","Data":"16e24156cb150c6e3db9cd76a8a264ed1454fe3279b2b439474c2e9c18185aec"} Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.130284 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.130258385 podStartE2EDuration="2.130258385s" podCreationTimestamp="2025-12-03 17:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:22:20.120097143 +0000 UTC m=+1595.321474024" watchObservedRunningTime="2025-12-03 17:22:20.130258385 +0000 UTC m=+1595.331635246" Dec 03 17:22:20 crc kubenswrapper[4758]: I1203 17:22:20.152305 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.152277546 podStartE2EDuration="2.152277546s" podCreationTimestamp="2025-12-03 17:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:22:20.137942852 +0000 UTC m=+1595.339319723" watchObservedRunningTime="2025-12-03 17:22:20.152277546 +0000 UTC m=+1595.353654407" Dec 03 17:22:21 crc kubenswrapper[4758]: I1203 17:22:21.125796 4758 generic.go:334] "Generic (PLEG): container finished" podID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerID="20d40fb3d8e847fd076300ae39ef23864f01d587d60b52073c6509f653a1882c" exitCode=0 Dec 03 17:22:21 crc kubenswrapper[4758]: I1203 17:22:21.135153 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60b25a93-398b-4e93-a0fd-cb366c2bc3b2","Type":"ContainerDied","Data":"20d40fb3d8e847fd076300ae39ef23864f01d587d60b52073c6509f653a1882c"} Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.075539 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.140136 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60b25a93-398b-4e93-a0fd-cb366c2bc3b2","Type":"ContainerDied","Data":"72e8e377ede390507999917e2a5eb8401ce5aa00bebd0d9e173825ab2ef75ba2"} Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.140212 4758 scope.go:117] "RemoveContainer" containerID="20d40fb3d8e847fd076300ae39ef23864f01d587d60b52073c6509f653a1882c" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.140243 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.189937 4758 scope.go:117] "RemoveContainer" containerID="f3395389398c63f9417e3a4f3653aa710b138bf479b204f563ee182a23f6d4c6" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.235277 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gsct\" (UniqueName: \"kubernetes.io/projected/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-kube-api-access-8gsct\") pod \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.235343 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-config-data\") pod \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.235539 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-combined-ca-bundle\") pod \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.235638 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-public-tls-certs\") pod \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.235722 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-logs\") pod \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.235918 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-internal-tls-certs\") pod \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\" (UID: \"60b25a93-398b-4e93-a0fd-cb366c2bc3b2\") " Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.238200 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-logs" (OuterVolumeSpecName: "logs") pod "60b25a93-398b-4e93-a0fd-cb366c2bc3b2" (UID: "60b25a93-398b-4e93-a0fd-cb366c2bc3b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.253064 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-kube-api-access-8gsct" (OuterVolumeSpecName: "kube-api-access-8gsct") pod "60b25a93-398b-4e93-a0fd-cb366c2bc3b2" (UID: "60b25a93-398b-4e93-a0fd-cb366c2bc3b2"). InnerVolumeSpecName "kube-api-access-8gsct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.299293 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "60b25a93-398b-4e93-a0fd-cb366c2bc3b2" (UID: "60b25a93-398b-4e93-a0fd-cb366c2bc3b2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.325974 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60b25a93-398b-4e93-a0fd-cb366c2bc3b2" (UID: "60b25a93-398b-4e93-a0fd-cb366c2bc3b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.336038 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "60b25a93-398b-4e93-a0fd-cb366c2bc3b2" (UID: "60b25a93-398b-4e93-a0fd-cb366c2bc3b2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.340578 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.340622 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.340634 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gsct\" (UniqueName: \"kubernetes.io/projected/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-kube-api-access-8gsct\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.340645 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.340655 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.372972 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-config-data" (OuterVolumeSpecName: "config-data") pod "60b25a93-398b-4e93-a0fd-cb366c2bc3b2" (UID: "60b25a93-398b-4e93-a0fd-cb366c2bc3b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.443199 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60b25a93-398b-4e93-a0fd-cb366c2bc3b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.491078 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.505347 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.523817 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:22 crc kubenswrapper[4758]: E1203 17:22:22.524916 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-api" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.524948 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-api" Dec 03 17:22:22 crc kubenswrapper[4758]: E1203 17:22:22.524996 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-log" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.525006 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-log" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.525236 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-log" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.525259 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" containerName="nova-api-api" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.527554 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.531600 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.531654 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.531840 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.553655 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.648440 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-public-tls-certs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.648504 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6swb6\" (UniqueName: \"kubernetes.io/projected/b461991e-58f6-4ff3-81d3-9cf2288d9b65-kube-api-access-6swb6\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.648636 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b461991e-58f6-4ff3-81d3-9cf2288d9b65-logs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.648727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-config-data\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.648783 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.648845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.750577 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.750818 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-public-tls-certs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.750850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6swb6\" (UniqueName: \"kubernetes.io/projected/b461991e-58f6-4ff3-81d3-9cf2288d9b65-kube-api-access-6swb6\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.750891 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b461991e-58f6-4ff3-81d3-9cf2288d9b65-logs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.750928 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-config-data\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.750975 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.752654 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b461991e-58f6-4ff3-81d3-9cf2288d9b65-logs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.759035 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-config-data\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.765069 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-public-tls-certs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.765070 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.772264 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6swb6\" (UniqueName: \"kubernetes.io/projected/b461991e-58f6-4ff3-81d3-9cf2288d9b65-kube-api-access-6swb6\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.777857 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " pod="openstack/nova-api-0" Dec 03 17:22:22 crc kubenswrapper[4758]: I1203 17:22:22.872478 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:22:23 crc kubenswrapper[4758]: I1203 17:22:23.138874 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b25a93-398b-4e93-a0fd-cb366c2bc3b2" path="/var/lib/kubelet/pods/60b25a93-398b-4e93-a0fd-cb366c2bc3b2/volumes" Dec 03 17:22:23 crc kubenswrapper[4758]: I1203 17:22:23.370056 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:22:23 crc kubenswrapper[4758]: I1203 17:22:23.621788 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 17:22:23 crc kubenswrapper[4758]: I1203 17:22:23.780952 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:22:23 crc kubenswrapper[4758]: I1203 17:22:23.781014 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:22:24 crc kubenswrapper[4758]: I1203 17:22:24.169510 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b461991e-58f6-4ff3-81d3-9cf2288d9b65","Type":"ContainerStarted","Data":"d0fc34ea02fb5712f5c26717176864428bc2e389b2d8605da62e87ecf5f8f023"} Dec 03 17:22:24 crc kubenswrapper[4758]: I1203 17:22:24.170046 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b461991e-58f6-4ff3-81d3-9cf2288d9b65","Type":"ContainerStarted","Data":"e1f514beb4453b1687bf12d6068c37dc748aad9f2f74d218b6b3104db6148ce3"} Dec 03 17:22:24 crc kubenswrapper[4758]: I1203 17:22:24.170068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b461991e-58f6-4ff3-81d3-9cf2288d9b65","Type":"ContainerStarted","Data":"ecbd0b931eee3f79ade53d3def7b3127f4c981c4b84b65ae184dfb6225502c64"} Dec 03 17:22:24 crc kubenswrapper[4758]: I1203 17:22:24.198898 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.198860714 podStartE2EDuration="2.198860714s" podCreationTimestamp="2025-12-03 17:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:22:24.189538275 +0000 UTC m=+1599.390915136" watchObservedRunningTime="2025-12-03 17:22:24.198860714 +0000 UTC m=+1599.400237575" Dec 03 17:22:28 crc kubenswrapper[4758]: I1203 17:22:28.622121 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 17:22:28 crc kubenswrapper[4758]: I1203 17:22:28.652918 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 17:22:28 crc kubenswrapper[4758]: I1203 17:22:28.780971 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:22:28 crc kubenswrapper[4758]: I1203 17:22:28.781033 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:22:29 crc kubenswrapper[4758]: I1203 17:22:29.287737 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 17:22:29 crc kubenswrapper[4758]: I1203 17:22:29.793858 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:22:29 crc kubenswrapper[4758]: I1203 17:22:29.793858 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:22:32 crc kubenswrapper[4758]: I1203 17:22:32.876015 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:22:32 crc kubenswrapper[4758]: I1203 17:22:32.877357 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:22:33 crc kubenswrapper[4758]: I1203 17:22:33.891203 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:22:33 crc kubenswrapper[4758]: I1203 17:22:33.893346 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:22:34 crc kubenswrapper[4758]: I1203 17:22:34.580480 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:22:38 crc kubenswrapper[4758]: I1203 17:22:38.787659 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:22:38 crc kubenswrapper[4758]: I1203 17:22:38.798542 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:22:38 crc kubenswrapper[4758]: I1203 17:22:38.799731 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:22:39 crc kubenswrapper[4758]: I1203 17:22:39.357521 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:22:41 crc kubenswrapper[4758]: I1203 17:22:41.394876 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:22:41 crc kubenswrapper[4758]: I1203 17:22:41.395439 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:22:41 crc kubenswrapper[4758]: I1203 17:22:41.395522 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:22:41 crc kubenswrapper[4758]: I1203 17:22:41.397337 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:22:41 crc kubenswrapper[4758]: I1203 17:22:41.397480 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" gracePeriod=600 Dec 03 17:22:41 crc kubenswrapper[4758]: E1203 17:22:41.522026 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.397727 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a"} Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.397805 4758 scope.go:117] "RemoveContainer" containerID="edeb1f97ccdd747653b82f053447c5a4a38ed821ec3f1cb524dc9338c014ae57" Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.397662 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" exitCode=0 Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.398725 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:22:42 crc kubenswrapper[4758]: E1203 17:22:42.398983 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.880275 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.881236 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.882617 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:22:42 crc kubenswrapper[4758]: I1203 17:22:42.889564 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:22:43 crc kubenswrapper[4758]: I1203 17:22:43.413924 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:22:43 crc kubenswrapper[4758]: I1203 17:22:43.420843 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:22:53 crc kubenswrapper[4758]: I1203 17:22:53.126419 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:22:53 crc kubenswrapper[4758]: E1203 17:22:53.127969 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.592277 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.593156 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" containerName="openstackclient" containerID="cri-o://5659d420ac31a290e23ff03a8dee2a674a17edcf17f81a32fdb9e77fc73d7428" gracePeriod=2 Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.623008 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.940264 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder14bb-account-delete-ctzv7"] Dec 03 17:23:02 crc kubenswrapper[4758]: E1203 17:23:02.940943 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" containerName="openstackclient" Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.940966 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" containerName="openstackclient" Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.941229 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" containerName="openstackclient" Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.942151 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.970300 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:23:02 crc kubenswrapper[4758]: I1203 17:23:02.995481 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder14bb-account-delete-ctzv7"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.131970 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hswgp\" (UniqueName: \"kubernetes.io/projected/3b4e949a-2e3c-49ff-ba65-f925a9a95336-kube-api-access-hswgp\") pod \"cinder14bb-account-delete-ctzv7\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.132245 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4e949a-2e3c-49ff-ba65-f925a9a95336-operator-scripts\") pod \"cinder14bb-account-delete-ctzv7\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.133333 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.133521 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data podName:070406cc-6d65-4fd0-81d7-264adead6bbd nodeName:}" failed. No retries permitted until 2025-12-03 17:23:03.633483313 +0000 UTC m=+1638.834860174 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data") pod "rabbitmq-server-0" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd") : configmap "rabbitmq-config-data" not found Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.172871 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement02ea-account-delete-zbwhh"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.179448 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron988a-account-delete-t55hx"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.181174 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.181665 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.212114 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement02ea-account-delete-zbwhh"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.240946 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4e949a-2e3c-49ff-ba65-f925a9a95336-operator-scripts\") pod \"cinder14bb-account-delete-ctzv7\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.241077 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hswgp\" (UniqueName: \"kubernetes.io/projected/3b4e949a-2e3c-49ff-ba65-f925a9a95336-kube-api-access-hswgp\") pod \"cinder14bb-account-delete-ctzv7\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.244211 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4e949a-2e3c-49ff-ba65-f925a9a95336-operator-scripts\") pod \"cinder14bb-account-delete-ctzv7\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.244585 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron988a-account-delete-t55hx"] Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.245254 4758 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.245384 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts podName:aac23ac3-72ea-49a7-9ded-6fabe28ff690 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:03.745361761 +0000 UTC m=+1638.946738622 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts") pod "ovn-controller-5czb6" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690") : configmap "ovncontroller-scripts" not found Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.283795 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.284448 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" containerID="cri-o://c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" gracePeriod=30 Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.285130 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="openstack-network-exporter" containerID="cri-o://43d7c9ec7f4e1c689ef29fc35a5e148631fe6789db20d2476a17525c48bb0c03" gracePeriod=30 Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.319079 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hswgp\" (UniqueName: \"kubernetes.io/projected/3b4e949a-2e3c-49ff-ba65-f925a9a95336-kube-api-access-hswgp\") pod \"cinder14bb-account-delete-ctzv7\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.345493 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" probeResult="failure" output=< Dec 03 17:23:03 crc kubenswrapper[4758]: 2025-12-03T17:23:03Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Dec 03 17:23:03 crc kubenswrapper[4758]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Dec 03 17:23:03 crc kubenswrapper[4758]: 2025-12-03T17:23:03Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Dec 03 17:23:03 crc kubenswrapper[4758]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Dec 03 17:23:03 crc kubenswrapper[4758]: > Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.352895 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde9ec06-c65f-45a5-9e90-5a12397f010e-operator-scripts\") pod \"neutron988a-account-delete-t55hx\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.353057 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7mbs\" (UniqueName: \"kubernetes.io/projected/014153c0-72a5-46ae-9065-ff34c290745a-kube-api-access-w7mbs\") pod \"placement02ea-account-delete-zbwhh\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.353134 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/014153c0-72a5-46ae-9065-ff34c290745a-operator-scripts\") pod \"placement02ea-account-delete-zbwhh\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.353481 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86r88\" (UniqueName: \"kubernetes.io/projected/bde9ec06-c65f-45a5-9e90-5a12397f010e-kube-api-access-86r88\") pod \"neutron988a-account-delete-t55hx\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.404773 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.457016 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86r88\" (UniqueName: \"kubernetes.io/projected/bde9ec06-c65f-45a5-9e90-5a12397f010e-kube-api-access-86r88\") pod \"neutron988a-account-delete-t55hx\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.457117 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde9ec06-c65f-45a5-9e90-5a12397f010e-operator-scripts\") pod \"neutron988a-account-delete-t55hx\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.457174 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7mbs\" (UniqueName: \"kubernetes.io/projected/014153c0-72a5-46ae-9065-ff34c290745a-kube-api-access-w7mbs\") pod \"placement02ea-account-delete-zbwhh\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.457202 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/014153c0-72a5-46ae-9065-ff34c290745a-operator-scripts\") pod \"placement02ea-account-delete-zbwhh\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.458066 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/014153c0-72a5-46ae-9065-ff34c290745a-operator-scripts\") pod \"placement02ea-account-delete-zbwhh\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.461520 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde9ec06-c65f-45a5-9e90-5a12397f010e-operator-scripts\") pod \"neutron988a-account-delete-t55hx\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.471779 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.473015 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="openstack-network-exporter" containerID="cri-o://da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a" gracePeriod=300 Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.525852 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicane6b7-account-delete-l2xbs"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.529350 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.562905 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicane6b7-account-delete-l2xbs"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.570587 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86r88\" (UniqueName: \"kubernetes.io/projected/bde9ec06-c65f-45a5-9e90-5a12397f010e-kube-api-access-86r88\") pod \"neutron988a-account-delete-t55hx\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.575118 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.575213 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data podName:e518390e-b4c4-49bb-913f-ddfc331456ab nodeName:}" failed. No retries permitted until 2025-12-03 17:23:04.075187308 +0000 UTC m=+1639.276564179 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data") pod "rabbitmq-cell1-server-0" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab") : configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.602613 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-6l8jp"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.604038 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.615785 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7mbs\" (UniqueName: \"kubernetes.io/projected/014153c0-72a5-46ae-9065-ff34c290745a-kube-api-access-w7mbs\") pod \"placement02ea-account-delete-zbwhh\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.646945 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-6l8jp"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.675330 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2q97\" (UniqueName: \"kubernetes.io/projected/cb717c57-e263-449f-896f-38707d61e72c-kube-api-access-t2q97\") pod \"barbicane6b7-account-delete-l2xbs\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.676130 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts\") pod \"barbicane6b7-account-delete-l2xbs\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.676332 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.676399 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data podName:070406cc-6d65-4fd0-81d7-264adead6bbd nodeName:}" failed. No retries permitted until 2025-12-03 17:23:04.67637463 +0000 UTC m=+1639.877751491 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data") pod "rabbitmq-server-0" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd") : configmap "rabbitmq-config-data" not found Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.708555 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4jmgb"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.780181 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts\") pod \"barbicane6b7-account-delete-l2xbs\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.780543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2q97\" (UniqueName: \"kubernetes.io/projected/cb717c57-e263-449f-896f-38707d61e72c-kube-api-access-t2q97\") pod \"barbicane6b7-account-delete-l2xbs\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.780848 4758 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 03 17:23:03 crc kubenswrapper[4758]: E1203 17:23:03.781029 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts podName:aac23ac3-72ea-49a7-9ded-6fabe28ff690 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:04.780939531 +0000 UTC m=+1639.982316392 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts") pod "ovn-controller-5czb6" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690") : configmap "ovncontroller-scripts" not found Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.781971 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0700b-account-delete-snfrl"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.783561 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.791854 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts\") pod \"barbicane6b7-account-delete-l2xbs\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.821453 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.919880 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2q97\" (UniqueName: \"kubernetes.io/projected/cb717c57-e263-449f-896f-38707d61e72c-kube-api-access-t2q97\") pod \"barbicane6b7-account-delete-l2xbs\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.928905 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4jmgb"] Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.936636 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="ovsdbserver-sb" containerID="cri-o://a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098" gracePeriod=300 Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.938621 4758 scope.go:117] "RemoveContainer" containerID="0808c7b33148d49a50727b199a344d2d5e5ae74f8fc8755e99598b3aaaffe6b6" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.945633 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.952925 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7znfc\" (UniqueName: \"kubernetes.io/projected/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-kube-api-access-7znfc\") pod \"novacell0700b-account-delete-snfrl\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:03 crc kubenswrapper[4758]: I1203 17:23:03.953118 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts\") pod \"novacell0700b-account-delete-snfrl\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.093617 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.110770 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0700b-account-delete-snfrl"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.113537 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts\") pod \"novacell0700b-account-delete-snfrl\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.114392 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7znfc\" (UniqueName: \"kubernetes.io/projected/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-kube-api-access-7znfc\") pod \"novacell0700b-account-delete-snfrl\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.115409 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.115557 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data podName:e518390e-b4c4-49bb-913f-ddfc331456ab nodeName:}" failed. No retries permitted until 2025-12-03 17:23:05.115524747 +0000 UTC m=+1640.316901608 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data") pod "rabbitmq-cell1-server-0" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab") : configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.115805 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.116370 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.132492 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts\") pod \"novacell0700b-account-delete-snfrl\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.279221 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7znfc\" (UniqueName: \"kubernetes.io/projected/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-kube-api-access-7znfc\") pod \"novacell0700b-account-delete-snfrl\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.325587 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-j7q7w"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.337240 4758 scope.go:117] "RemoveContainer" containerID="fe0431b466f23c5236045a7abb2108a2f3b4390cfee17ffd736b71cb61050e03" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.373139 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-j7q7w"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.405702 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-q2vlf"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.415026 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-q2vlf"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.436245 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell1a6be-account-delete-tcd65"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.438087 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.460212 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovsdbserver-sb-0" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="ovsdbserver-sb" probeResult="failure" output="" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.512818 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell1a6be-account-delete-tcd65"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.534333 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-55pgs"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.534819 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-55pgs" podUID="2923bcc5-72ec-493c-ad10-9f171801ae4a" containerName="openstack-network-exporter" containerID="cri-o://5877538da9e694dfac7b157eb99747894a3276ae973e31c6d9517ea26fd87fd7" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.542621 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.549854 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5czb6"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.582797 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.582948 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn6p7\" (UniqueName: \"kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.585337 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-t442n"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.602898 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapic90a-account-delete-gg2x2"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.605756 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.616204 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-t442n"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.635658 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-w5n6b"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.650332 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-5mttb"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.660348 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic90a-account-delete-gg2x2"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.673937 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-w5n6b"] Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.691574 4758 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.691672 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:05.191646114 +0000 UTC m=+1640.393022975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : configmap "openstack-cell1-scripts" not found Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.698337 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.699803 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="openstack-network-exporter" containerID="cri-o://ec3db493149a77626469cdaa439ac0d0670962001d5407b77c4d6b653c833931" gracePeriod=300 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.700742 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.701336 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn6p7\" (UniqueName: \"kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.702110 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.702202 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data podName:070406cc-6d65-4fd0-81d7-264adead6bbd nodeName:}" failed. No retries permitted until 2025-12-03 17:23:06.702172436 +0000 UTC m=+1641.903549307 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data") pod "rabbitmq-server-0" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd") : configmap "rabbitmq-config-data" not found Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.707817 4758 projected.go:194] Error preparing data for projected volume kube-api-access-mn6p7 for pod openstack/novacell1a6be-account-delete-tcd65: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.707899 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7 podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:05.207877779 +0000 UTC m=+1640.409254640 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mn6p7" (UniqueName: "kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.715096 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.715386 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="cinder-scheduler" containerID="cri-o://df8bc0faf1cabc23c62d9aea7423ca778a9372ef7f9e311140e90eae1b53b014" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.715918 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="probe" containerID="cri-o://a8202adddc1c7cf7ab92d5ed5e61b0495c0d800596e0b0472d95b872e5eca55f" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.804697 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-qr4sd"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.812439 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wctjz\" (UniqueName: \"kubernetes.io/projected/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-kube-api-access-wctjz\") pod \"novaapic90a-account-delete-gg2x2\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.812557 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts\") pod \"novaapic90a-account-delete-gg2x2\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.812748 4758 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 03 17:23:04 crc kubenswrapper[4758]: E1203 17:23:04.812814 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts podName:aac23ac3-72ea-49a7-9ded-6fabe28ff690 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:06.8127931 +0000 UTC m=+1642.014169961 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts") pod "ovn-controller-5czb6" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690") : configmap "ovncontroller-scripts" not found Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.838182 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-qr4sd"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.864827 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancee859-account-delete-bb4lj"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.881024 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="ovsdbserver-nb" containerID="cri-o://66ede82dc9e0fb893690735dd61d3b58c7e3794bb7c61e6dce24226c9e1b2e55" gracePeriod=300 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.888229 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69dc74dcbf-wzmb5"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.889145 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69dc74dcbf-wzmb5" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-api" containerID="cri-o://2496f74c6aa8a3c995f6fcb782d022aa7cb8ab73e8224330eecc746a01760d7c" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.890275 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.890482 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69dc74dcbf-wzmb5" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-httpd" containerID="cri-o://6637a67a10c678e5c8c666412014de74c5391a41c691d978da31902186b3e7c0" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.900989 4758 generic.go:334] "Generic (PLEG): container finished" podID="daf64139-0696-484a-870f-023898fbeb08" containerID="43d7c9ec7f4e1c689ef29fc35a5e148631fe6789db20d2476a17525c48bb0c03" exitCode=2 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.901084 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"daf64139-0696-484a-870f-023898fbeb08","Type":"ContainerDied","Data":"43d7c9ec7f4e1c689ef29fc35a5e148631fe6789db20d2476a17525c48bb0c03"} Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.910349 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-55pgs_2923bcc5-72ec-493c-ad10-9f171801ae4a/openstack-network-exporter/0.log" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.910731 4758 generic.go:334] "Generic (PLEG): container finished" podID="2923bcc5-72ec-493c-ad10-9f171801ae4a" containerID="5877538da9e694dfac7b157eb99747894a3276ae973e31c6d9517ea26fd87fd7" exitCode=2 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.910806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-55pgs" event={"ID":"2923bcc5-72ec-493c-ad10-9f171801ae4a","Type":"ContainerDied","Data":"5877538da9e694dfac7b157eb99747894a3276ae973e31c6d9517ea26fd87fd7"} Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.932037 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts\") pod \"glancee859-account-delete-bb4lj\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.932334 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts\") pod \"novaapic90a-account-delete-gg2x2\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.932647 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6skc\" (UniqueName: \"kubernetes.io/projected/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-kube-api-access-p6skc\") pod \"glancee859-account-delete-bb4lj\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.934086 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts\") pod \"novaapic90a-account-delete-gg2x2\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.934400 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-68d78d4d4d-rhzqc"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.934639 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-68d78d4d4d-rhzqc" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-log" containerID="cri-o://7ff34332c6d8c7344235f023f03dbde628713fc91c197634ffc70e74d0fc3ae7" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.935157 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-68d78d4d4d-rhzqc" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-api" containerID="cri-o://e9526364cd771e1d73395320b28267b014ffd81941f74ef5e036bbd181871def" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.942593 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wctjz\" (UniqueName: \"kubernetes.io/projected/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-kube-api-access-wctjz\") pod \"novaapic90a-account-delete-gg2x2\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.961093 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7a370910-1957-4bd8-891b-7f1065c5e286/ovsdbserver-sb/0.log" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.961171 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a370910-1957-4bd8-891b-7f1065c5e286" containerID="da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a" exitCode=2 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.961198 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a370910-1957-4bd8-891b-7f1065c5e286" containerID="a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098" exitCode=143 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.961235 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7a370910-1957-4bd8-891b-7f1065c5e286","Type":"ContainerDied","Data":"da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a"} Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.961314 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7a370910-1957-4bd8-891b-7f1065c5e286","Type":"ContainerDied","Data":"a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098"} Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.974295 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wctjz\" (UniqueName: \"kubernetes.io/projected/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-kube-api-access-wctjz\") pod \"novaapic90a-account-delete-gg2x2\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.979807 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.980138 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api-log" containerID="cri-o://b515931c8d15bc7880738113558415685d689f0ef349e841890a0dc9c902625e" gracePeriod=30 Dec 03 17:23:04 crc kubenswrapper[4758]: I1203 17:23:04.980828 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api" containerID="cri-o://ef163a1a1695789c803e3b2391ac85959aee35c4dc735cf405bb6587fe19ed5f" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.002886 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancee859-account-delete-bb4lj"] Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.049131 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts\") pod \"glancee859-account-delete-bb4lj\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.049286 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6skc\" (UniqueName: \"kubernetes.io/projected/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-kube-api-access-p6skc\") pod \"glancee859-account-delete-bb4lj\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.050639 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts\") pod \"glancee859-account-delete-bb4lj\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.050758 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-s582w"] Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.050987 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerName="dnsmasq-dns" containerID="cri-o://3c175f3c3abbd9f78b887a0d32c0fdf73322a6b77388cdd343be7a3ed9da0d73" gracePeriod=10 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.079650 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.080590 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-server" containerID="cri-o://def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.082325 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="swift-recon-cron" containerID="cri-o://1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.082434 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="rsync" containerID="cri-o://fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.082500 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-expirer" containerID="cri-o://1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.082568 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-updater" containerID="cri-o://1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.082606 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-auditor" containerID="cri-o://187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.082982 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-replicator" containerID="cri-o://98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083064 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-server" containerID="cri-o://5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083147 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-updater" containerID="cri-o://a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083222 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-auditor" containerID="cri-o://4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083287 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-replicator" containerID="cri-o://48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083332 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-server" containerID="cri-o://dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083419 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-reaper" containerID="cri-o://7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.083485 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-auditor" containerID="cri-o://8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.084791 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-replicator" containerID="cri-o://93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.135864 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6skc\" (UniqueName: \"kubernetes.io/projected/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-kube-api-access-p6skc\") pod \"glancee859-account-delete-bb4lj\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:05 crc kubenswrapper[4758]: E1203 17:23:05.151559 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:05 crc kubenswrapper[4758]: E1203 17:23:05.151650 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data podName:e518390e-b4c4-49bb-913f-ddfc331456ab nodeName:}" failed. No retries permitted until 2025-12-03 17:23:07.151627639 +0000 UTC m=+1642.353004500 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data") pod "rabbitmq-cell1-server-0" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab") : configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.235157 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502df21c-0370-40b8-b3ef-0e55992b3008" path="/var/lib/kubelet/pods/502df21c-0370-40b8-b3ef-0e55992b3008/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.238122 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b98f774-a8d0-460e-a718-c119dbd1f203" path="/var/lib/kubelet/pods/6b98f774-a8d0-460e-a718-c119dbd1f203/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.271531 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1f01fd7-322b-4576-86c8-83f8a06a7792" path="/var/lib/kubelet/pods/a1f01fd7-322b-4576-86c8-83f8a06a7792/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.273172 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbfaa3a1-7e61-4537-b953-07269f386adc" path="/var/lib/kubelet/pods/dbfaa3a1-7e61-4537-b953-07269f386adc/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.287098 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6bb6321-e853-40ce-a5f6-7b7a53e26595" path="/var/lib/kubelet/pods/e6bb6321-e853-40ce-a5f6-7b7a53e26595/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.287837 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb23ba13-deef-46a6-a81d-a223ad56024b" path="/var/lib/kubelet/pods/eb23ba13-deef-46a6-a81d-a223ad56024b/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.293271 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe2b439-b61f-4a3e-accc-abc261fc216f" path="/var/lib/kubelet/pods/efe2b439-b61f-4a3e-accc-abc261fc216f/volumes" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.301571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.297751 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-48jkw"] Dec 03 17:23:05 crc kubenswrapper[4758]: E1203 17:23:05.308016 4758 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 03 17:23:05 crc kubenswrapper[4758]: E1203 17:23:05.308186 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:06.308152353 +0000 UTC m=+1641.509529214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : configmap "openstack-cell1-scripts" not found Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.309460 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn6p7\" (UniqueName: \"kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.316077 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-48jkw"] Dec 03 17:23:05 crc kubenswrapper[4758]: E1203 17:23:05.344604 4758 projected.go:194] Error preparing data for projected volume kube-api-access-mn6p7 for pod openstack/novacell1a6be-account-delete-tcd65: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:05 crc kubenswrapper[4758]: E1203 17:23:05.344744 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7 podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:06.344715463 +0000 UTC m=+1641.546092324 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-mn6p7" (UniqueName: "kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.452659 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.485298 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" containerID="cri-o://ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.485896 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" containerID="cri-o://e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.504617 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.505134 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-log" containerID="cri-o://39dfeec538bbe7eb43d14ac4885d21560e65527c9acf6664be573a26763ad32b" gracePeriod=30 Dec 03 17:23:05 crc kubenswrapper[4758]: I1203 17:23:05.507709 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-metadata" containerID="cri-o://08867454fb2022e6047c48092dc940ce2f60d0e3bb28b9647d08e3cc414e048e" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.548137 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.558963 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.559251 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-log" containerID="cri-o://aacdb99dee5e45686e33def0624769051138394c943550e16068c2601899b7a4" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.559918 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-httpd" containerID="cri-o://b226eaf6e02f5bf23bb058d2357c7eb31da688995dc12dbdc9fe34fe8dbc7091" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.569967 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7bb9966db6-phk6t"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.570517 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener-log" containerID="cri-o://9c02a3775a91ddea9562f2c00063105c43fac3fa4443572b50d1f8eb0a7c2803" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.572403 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener" containerID="cri-o://efb795a450acef7b396db339028440c950cbfa2f32b9d7f9c051e83e71c85fb6" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.621875 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.622509 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-api" containerID="cri-o://d0fc34ea02fb5712f5c26717176864428bc2e389b2d8605da62e87ecf5f8f023" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.622754 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-log" containerID="cri-o://e1f514beb4453b1687bf12d6068c37dc748aad9f2f74d218b6b3104db6148ce3" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.691572 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.692070 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-log" containerID="cri-o://b86b8be6d1e76610577e86e34e754975aa88fa3ee8f7d13944550ef42f337ec8" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.692976 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-httpd" containerID="cri-o://9d2c93db5d7424078ddb1bb905ea651dea2607f1f94df3e14e8eaf9f6eec920f" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.743360 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-57b5ff9967-k6s52"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.744003 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-57b5ff9967-k6s52" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker-log" containerID="cri-o://7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.745460 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-57b5ff9967-k6s52" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker" containerID="cri-o://f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.839406 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-576b5fdbd4-qmczw"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.839812 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-576b5fdbd4-qmczw" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api-log" containerID="cri-o://99106e146935a884c8eae3a30d8b26d2549f367f4b50ca0f4536926fc866b59a" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.840511 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-576b5fdbd4-qmczw" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api" containerID="cri-o://c1e072acd4c62b59d1cb154ec06ad176f5ed4d1bdd3cc652119c966d7118b7f9" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.879306 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xxw46"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.913849 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xxw46"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.981739 4758 generic.go:334] "Generic (PLEG): container finished" podID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerID="6637a67a10c678e5c8c666412014de74c5391a41c691d978da31902186b3e7c0" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.981819 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dc74dcbf-wzmb5" event={"ID":"bd709657-fb8d-4c81-873f-a0b44d63abef","Type":"ContainerDied","Data":"6637a67a10c678e5c8c666412014de74c5391a41c691d978da31902186b3e7c0"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.985466 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d870c33d-67ba-4991-8624-4d0a28cb8542/ovsdbserver-nb/0.log" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.985540 4758 generic.go:334] "Generic (PLEG): container finished" podID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerID="ec3db493149a77626469cdaa439ac0d0670962001d5407b77c4d6b653c833931" exitCode=2 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.985570 4758 generic.go:334] "Generic (PLEG): container finished" podID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerID="66ede82dc9e0fb893690735dd61d3b58c7e3794bb7c61e6dce24226c9e1b2e55" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.985642 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d870c33d-67ba-4991-8624-4d0a28cb8542","Type":"ContainerDied","Data":"ec3db493149a77626469cdaa439ac0d0670962001d5407b77c4d6b653c833931"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.985706 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d870c33d-67ba-4991-8624-4d0a28cb8542","Type":"ContainerDied","Data":"66ede82dc9e0fb893690735dd61d3b58c7e3794bb7c61e6dce24226c9e1b2e55"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.988032 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:05.988454 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6df131e1-f13f-4943-95a0-18086525f883" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6bf75a7b95cf3a4d7d3c7984e0780c10a5e1cb76fbe0816ad9b844b0ed2b0584" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.005124 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.017886 4758 generic.go:334] "Generic (PLEG): container finished" podID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerID="b515931c8d15bc7880738113558415685d689f0ef349e841890a0dc9c902625e" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.018389 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b","Type":"ContainerDied","Data":"b515931c8d15bc7880738113558415685d689f0ef349e841890a0dc9c902625e"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.023061 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1a6be-account-delete-tcd65"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.032127 4758 generic.go:334] "Generic (PLEG): container finished" podID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerID="aacdb99dee5e45686e33def0624769051138394c943550e16068c2601899b7a4" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.032347 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b49b1d3-cbb4-49da-8244-69e15595da33","Type":"ContainerDied","Data":"aacdb99dee5e45686e33def0624769051138394c943550e16068c2601899b7a4"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.063396 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-a6be-account-create-update-vfdk6"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.067951 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.067985 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.067994 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068032 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068041 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068049 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068056 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068065 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068072 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068079 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068087 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068093 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068155 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068193 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068206 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068228 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068239 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068257 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068266 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068294 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.068305 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.073277 4758 generic.go:334] "Generic (PLEG): container finished" podID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.073344 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerDied","Data":"ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.076797 4758 generic.go:334] "Generic (PLEG): container finished" podID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerID="3c175f3c3abbd9f78b887a0d32c0fdf73322a6b77388cdd343be7a3ed9da0d73" exitCode=0 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.076934 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" event={"ID":"091fea73-c280-4583-810c-1cfe5f1f0fe1","Type":"ContainerDied","Data":"3c175f3c3abbd9f78b887a0d32c0fdf73322a6b77388cdd343be7a3ed9da0d73"} Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.089638 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a370910_1957_4bd8_891b_7f1065c5e286.slice/crio-da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a370910_1957_4bd8_891b_7f1065c5e286.slice/crio-conmon-da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946a9b5f_7abf_4b3e_bc51_08a45e73ff30.slice/crio-conmon-1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd50c7a0d_80f6_4c8c_a516_29dbe3a6397b.slice/crio-b515931c8d15bc7880738113558415685d689f0ef349e841890a0dc9c902625e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946a9b5f_7abf_4b3e_bc51_08a45e73ff30.slice/crio-conmon-187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946a9b5f_7abf_4b3e_bc51_08a45e73ff30.slice/crio-conmon-93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946a9b5f_7abf_4b3e_bc51_08a45e73ff30.slice/crio-98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f328daf_ac93_432b_8645_14b7030a24d1.slice/crio-7ff34332c6d8c7344235f023f03dbde628713fc91c197634ffc70e74d0fc3ae7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod091fea73_c280_4583_810c_1cfe5f1f0fe1.slice/crio-3c175f3c3abbd9f78b887a0d32c0fdf73322a6b77388cdd343be7a3ed9da0d73.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.092793 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerName="rabbitmq" containerID="cri-o://aa23c937b63cdc7de84a3a65b1ab3cb567e5f8d745fd98201fbefbe82494cbc1" gracePeriod=604800 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.101031 4758 generic.go:334] "Generic (PLEG): container finished" podID="84469de1-c706-4442-8610-31188889dad1" containerID="39dfeec538bbe7eb43d14ac4885d21560e65527c9acf6664be573a26763ad32b" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.101157 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84469de1-c706-4442-8610-31188889dad1","Type":"ContainerDied","Data":"39dfeec538bbe7eb43d14ac4885d21560e65527c9acf6664be573a26763ad32b"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.127959 4758 generic.go:334] "Generic (PLEG): container finished" podID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerID="99106e146935a884c8eae3a30d8b26d2549f367f4b50ca0f4536926fc866b59a" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: W1203 17:23:06.127995 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbde9ec06_c65f_45a5_9e90_5a12397f010e.slice/crio-89739162cf262cead564f1509b480ac13e65ee0c4239567ee0aee5f6c98a8e72 WatchSource:0}: Error finding container 89739162cf262cead564f1509b480ac13e65ee0c4239567ee0aee5f6c98a8e72: Status 404 returned error can't find the container with id 89739162cf262cead564f1509b480ac13e65ee0c4239567ee0aee5f6c98a8e72 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.128032 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-576b5fdbd4-qmczw" event={"ID":"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665","Type":"ContainerDied","Data":"99106e146935a884c8eae3a30d8b26d2549f367f4b50ca0f4536926fc866b59a"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.141201 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-a6be-account-create-update-vfdk6"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.169315 4758 generic.go:334] "Generic (PLEG): container finished" podID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerID="b86b8be6d1e76610577e86e34e754975aa88fa3ee8f7d13944550ef42f337ec8" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.169446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be629420-0e78-44d4-9ca1-4ee1601d4ca6","Type":"ContainerDied","Data":"b86b8be6d1e76610577e86e34e754975aa88fa3ee8f7d13944550ef42f337ec8"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.185115 4758 generic.go:334] "Generic (PLEG): container finished" podID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerID="9c02a3775a91ddea9562f2c00063105c43fac3fa4443572b50d1f8eb0a7c2803" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.185214 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" event={"ID":"53b74dc9-5836-4c5c-bae8-803f64c9de12","Type":"ContainerDied","Data":"9c02a3775a91ddea9562f2c00063105c43fac3fa4443572b50d1f8eb0a7c2803"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.243207 4758 generic.go:334] "Generic (PLEG): container finished" podID="1f328daf-ac93-432b-8645-14b7030a24d1" containerID="7ff34332c6d8c7344235f023f03dbde628713fc91c197634ffc70e74d0fc3ae7" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.243409 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d78d4d4d-rhzqc" event={"ID":"1f328daf-ac93-432b-8645-14b7030a24d1","Type":"ContainerDied","Data":"7ff34332c6d8c7344235f023f03dbde628713fc91c197634ffc70e74d0fc3ae7"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.246937 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerName="rabbitmq" containerID="cri-o://ffded9f7e8bb729a9659af5430bde139dac85bb5e31acabdfd96c07c75f4d43d" gracePeriod=604800 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.256439 4758 generic.go:334] "Generic (PLEG): container finished" podID="790de883-aaa9-4faa-b18c-61ef4d093ab4" containerID="5659d420ac31a290e23ff03a8dee2a674a17edcf17f81a32fdb9e77fc73d7428" exitCode=137 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.265289 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron988a-account-delete-t55hx"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.275132 4758 generic.go:334] "Generic (PLEG): container finished" podID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerID="e1f514beb4453b1687bf12d6068c37dc748aad9f2f74d218b6b3104db6148ce3" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.275771 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b461991e-58f6-4ff3-81d3-9cf2288d9b65","Type":"ContainerDied","Data":"e1f514beb4453b1687bf12d6068c37dc748aad9f2f74d218b6b3104db6148ce3"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.277081 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement02ea-account-delete-zbwhh"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.284951 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlkz4"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.293449 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.293697 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerName="nova-cell1-conductor-conductor" containerID="cri-o://c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.319222 4758 generic.go:334] "Generic (PLEG): container finished" podID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerID="7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4" exitCode=143 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.319292 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57b5ff9967-k6s52" event={"ID":"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c","Type":"ContainerDied","Data":"7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4"} Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.325512 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlkz4"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.337859 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.338140 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="ebee961b-48e5-4983-bb45-c14bbcc164f8" containerName="nova-cell0-conductor-conductor" containerID="cri-o://b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.352665 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-mn6p7 operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell1a6be-account-delete-tcd65" podUID="154d62cd-e6fa-4b15-ab82-7f486ea22eaa" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.359814 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9tms7"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.386528 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9tms7"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.386614 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.386874 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" containerName="nova-scheduler-scheduler" containerID="cri-o://0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.389436 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098 is running failed: container process not found" containerID="a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.399830 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.400099 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn6p7\" (UniqueName: \"kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.400754 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098 is running failed: container process not found" containerID="a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.401245 4758 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.401308 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:08.401289334 +0000 UTC m=+1643.602666195 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : configmap "openstack-cell1-scripts" not found Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.405879 4758 projected.go:194] Error preparing data for projected volume kube-api-access-mn6p7 for pod openstack/novacell1a6be-account-delete-tcd65: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.405986 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7 podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:08.405958719 +0000 UTC m=+1643.607335760 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-mn6p7" (UniqueName: "kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.406045 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098 is running failed: container process not found" containerID="a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.406105 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="ovsdbserver-sb" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.406106 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.438637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.457178 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7a370910-1957-4bd8-891b-7f1065c5e286/ovsdbserver-sb/0.log" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.457267 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.497406 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.506278 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.515945 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.516013 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerName="nova-cell1-conductor-conductor" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.534439 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="galera" containerID="cri-o://d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42" gracePeriod=30 Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623073 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-metrics-certs-tls-certs\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623322 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwlhn\" (UniqueName: \"kubernetes.io/projected/7a370910-1957-4bd8-891b-7f1065c5e286-kube-api-access-rwlhn\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623489 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdb-rundir\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623611 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdbserver-sb-tls-certs\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-config\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623743 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-scripts\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623763 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.623789 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-combined-ca-bundle\") pod \"7a370910-1957-4bd8-891b-7f1065c5e286\" (UID: \"7a370910-1957-4bd8-891b-7f1065c5e286\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.624419 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.625123 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-config" (OuterVolumeSpecName: "config") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.625217 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.626190 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-scripts" (OuterVolumeSpecName: "scripts") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.640279 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a370910-1957-4bd8-891b-7f1065c5e286-kube-api-access-rwlhn" (OuterVolumeSpecName: "kube-api-access-rwlhn") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "kube-api-access-rwlhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.665376 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.752149 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.752235 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.752247 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data podName:070406cc-6d65-4fd0-81d7-264adead6bbd nodeName:}" failed. No retries permitted until 2025-12-03 17:23:10.752219587 +0000 UTC m=+1645.953596448 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data") pod "rabbitmq-server-0" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd") : configmap "rabbitmq-config-data" not found Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.752444 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a370910-1957-4bd8-891b-7f1065c5e286-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.752527 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.752546 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwlhn\" (UniqueName: \"kubernetes.io/projected/7a370910-1957-4bd8-891b-7f1065c5e286-kube-api-access-rwlhn\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.801411 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicane6b7-account-delete-l2xbs"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.806967 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder14bb-account-delete-ctzv7"] Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.810661 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.814533 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.875109 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.875198 4758 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.875218 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:06 crc kubenswrapper[4758]: E1203 17:23:06.875282 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts podName:aac23ac3-72ea-49a7-9ded-6fabe28ff690 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:10.875262974 +0000 UTC m=+1646.076639835 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts") pod "ovn-controller-5czb6" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690") : configmap "ovncontroller-scripts" not found Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.886845 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.923696 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-55pgs_2923bcc5-72ec-493c-ad10-9f171801ae4a/openstack-network-exporter/0.log" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.923807 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.992927 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-metrics-certs-tls-certs\") pod \"2923bcc5-72ec-493c-ad10-9f171801ae4a\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993033 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-combined-ca-bundle\") pod \"2923bcc5-72ec-493c-ad10-9f171801ae4a\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993068 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-sb\") pod \"091fea73-c280-4583-810c-1cfe5f1f0fe1\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993394 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l58rt\" (UniqueName: \"kubernetes.io/projected/091fea73-c280-4583-810c-1cfe5f1f0fe1-kube-api-access-l58rt\") pod \"091fea73-c280-4583-810c-1cfe5f1f0fe1\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993436 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pz2f\" (UniqueName: \"kubernetes.io/projected/2923bcc5-72ec-493c-ad10-9f171801ae4a-kube-api-access-8pz2f\") pod \"2923bcc5-72ec-493c-ad10-9f171801ae4a\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993532 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovs-rundir\") pod \"2923bcc5-72ec-493c-ad10-9f171801ae4a\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993556 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-nb\") pod \"091fea73-c280-4583-810c-1cfe5f1f0fe1\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993620 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovn-rundir\") pod \"2923bcc5-72ec-493c-ad10-9f171801ae4a\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993704 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2923bcc5-72ec-493c-ad10-9f171801ae4a-config\") pod \"2923bcc5-72ec-493c-ad10-9f171801ae4a\" (UID: \"2923bcc5-72ec-493c-ad10-9f171801ae4a\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993745 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-svc\") pod \"091fea73-c280-4583-810c-1cfe5f1f0fe1\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993781 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-swift-storage-0\") pod \"091fea73-c280-4583-810c-1cfe5f1f0fe1\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.993813 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-config\") pod \"091fea73-c280-4583-810c-1cfe5f1f0fe1\" (UID: \"091fea73-c280-4583-810c-1cfe5f1f0fe1\") " Dec 03 17:23:06 crc kubenswrapper[4758]: I1203 17:23:06.996993 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "2923bcc5-72ec-493c-ad10-9f171801ae4a" (UID: "2923bcc5-72ec-493c-ad10-9f171801ae4a"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.002263 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "2923bcc5-72ec-493c-ad10-9f171801ae4a" (UID: "2923bcc5-72ec-493c-ad10-9f171801ae4a"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.003671 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2923bcc5-72ec-493c-ad10-9f171801ae4a-config" (OuterVolumeSpecName: "config") pod "2923bcc5-72ec-493c-ad10-9f171801ae4a" (UID: "2923bcc5-72ec-493c-ad10-9f171801ae4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.004336 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d870c33d-67ba-4991-8624-4d0a28cb8542/ovsdbserver-nb/0.log" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.004426 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.020968 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.023818 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0700b-account-delete-snfrl"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.038926 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.056123 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2923bcc5-72ec-493c-ad10-9f171801ae4a-kube-api-access-8pz2f" (OuterVolumeSpecName: "kube-api-access-8pz2f") pod "2923bcc5-72ec-493c-ad10-9f171801ae4a" (UID: "2923bcc5-72ec-493c-ad10-9f171801ae4a"). InnerVolumeSpecName "kube-api-access-8pz2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.106219 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pz2f\" (UniqueName: \"kubernetes.io/projected/2923bcc5-72ec-493c-ad10-9f171801ae4a-kube-api-access-8pz2f\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.106258 4758 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.106267 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2923bcc5-72ec-493c-ad10-9f171801ae4a-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.106276 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2923bcc5-72ec-493c-ad10-9f171801ae4a-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.106287 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.129012 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="6df131e1-f13f-4943-95a0-18086525f883" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.199:6080/vnc_lite.html\": dial tcp 10.217.0.199:6080: connect: connection refused" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.129669 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091fea73-c280-4583-810c-1cfe5f1f0fe1-kube-api-access-l58rt" (OuterVolumeSpecName: "kube-api-access-l58rt") pod "091fea73-c280-4583-810c-1cfe5f1f0fe1" (UID: "091fea73-c280-4583-810c-1cfe5f1f0fe1"). InnerVolumeSpecName "kube-api-access-l58rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.129836 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207210 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-combined-ca-bundle\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207364 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdb-rundir\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207414 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-scripts\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207474 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-metrics-certs-tls-certs\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207531 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nfkk\" (UniqueName: \"kubernetes.io/projected/790de883-aaa9-4faa-b18c-61ef4d093ab4-kube-api-access-9nfkk\") pod \"790de883-aaa9-4faa-b18c-61ef4d093ab4\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207613 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config\") pod \"790de883-aaa9-4faa-b18c-61ef4d093ab4\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207655 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdbserver-nb-tls-certs\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207847 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-config\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207887 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-combined-ca-bundle\") pod \"790de883-aaa9-4faa-b18c-61ef4d093ab4\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207955 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pftgw\" (UniqueName: \"kubernetes.io/projected/d870c33d-67ba-4991-8624-4d0a28cb8542-kube-api-access-pftgw\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.207976 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config-secret\") pod \"790de883-aaa9-4faa-b18c-61ef4d093ab4\" (UID: \"790de883-aaa9-4faa-b18c-61ef4d093ab4\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.208020 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d870c33d-67ba-4991-8624-4d0a28cb8542\" (UID: \"d870c33d-67ba-4991-8624-4d0a28cb8542\") " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.208572 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l58rt\" (UniqueName: \"kubernetes.io/projected/091fea73-c280-4583-810c-1cfe5f1f0fe1-kube-api-access-l58rt\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.208797 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.208871 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data podName:e518390e-b4c4-49bb-913f-ddfc331456ab nodeName:}" failed. No retries permitted until 2025-12-03 17:23:11.208850012 +0000 UTC m=+1646.410226873 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data") pod "rabbitmq-cell1-server-0" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab") : configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.212639 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-config" (OuterVolumeSpecName: "config") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.215325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-scripts" (OuterVolumeSpecName: "scripts") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.221334 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.221832 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.249176 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.249268 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="galera" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.284966 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790de883-aaa9-4faa-b18c-61ef4d093ab4-kube-api-access-9nfkk" (OuterVolumeSpecName: "kube-api-access-9nfkk") pod "790de883-aaa9-4faa-b18c-61ef4d093ab4" (UID: "790de883-aaa9-4faa-b18c-61ef4d093ab4"). InnerVolumeSpecName "kube-api-access-9nfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.307233 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.310577 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nfkk\" (UniqueName: \"kubernetes.io/projected/790de883-aaa9-4faa-b18c-61ef4d093ab4-kube-api-access-9nfkk\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.310607 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.310632 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.310643 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.310657 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d870c33d-67ba-4991-8624-4d0a28cb8542-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.319991 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52cd754b-1015-487b-a51c-4d7117cfae33" path="/var/lib/kubelet/pods/52cd754b-1015-487b-a51c-4d7117cfae33/volumes" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.322173 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "7a370910-1957-4bd8-891b-7f1065c5e286" (UID: "7a370910-1957-4bd8-891b-7f1065c5e286"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.324831 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f" path="/var/lib/kubelet/pods/6dea5ef6-5ae7-4aff-a453-dbdcfea7ea6f/volumes" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.325898 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96f73920-eaca-458d-8b54-375f8c0928a3" path="/var/lib/kubelet/pods/96f73920-eaca-458d-8b54-375f8c0928a3/volumes" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.332457 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d870c33d-67ba-4991-8624-4d0a28cb8542-kube-api-access-pftgw" (OuterVolumeSpecName: "kube-api-access-pftgw") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "kube-api-access-pftgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.334730 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1bc708e-493d-4597-9a37-501be2e8641a" path="/var/lib/kubelet/pods/e1bc708e-493d-4597-9a37-501be2e8641a/volumes" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.335613 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f24e819a-1dee-418a-99cd-885cf9289b48" path="/var/lib/kubelet/pods/f24e819a-1dee-418a-99cd-885cf9289b48/volumes" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.344502 4758 generic.go:334] "Generic (PLEG): container finished" podID="6df131e1-f13f-4943-95a0-18086525f883" containerID="6bf75a7b95cf3a4d7d3c7984e0780c10a5e1cb76fbe0816ad9b844b0ed2b0584" exitCode=0 Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.353908 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6fc8f59f6f-7zxl4"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.353948 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6df131e1-f13f-4943-95a0-18086525f883","Type":"ContainerDied","Data":"6bf75a7b95cf3a4d7d3c7984e0780c10a5e1cb76fbe0816ad9b844b0ed2b0584"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.353971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement02ea-account-delete-zbwhh" event={"ID":"014153c0-72a5-46ae-9065-ff34c290745a","Type":"ContainerStarted","Data":"484b8a0e84e8e29e4a83fcb1f26cd7fc46d4f41e4674efa1174c310f89876926"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.353983 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement02ea-account-delete-zbwhh" event={"ID":"014153c0-72a5-46ae-9065-ff34c290745a","Type":"ContainerStarted","Data":"020331e515ccd2014f34f68d90c60679466b1826ddc48a69b7e2e4da84f2079e"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.354181 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-httpd" containerID="cri-o://033147cfa4974a6e9865e1dce4a3ab80379ee25004c515e9a54cca5bc2a76e0e" gracePeriod=30 Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.354287 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-server" containerID="cri-o://6b35ae8e6d0ff4c3cd07edce5d23b46583d8b4af2ecf93f7e9477313990aa41c" gracePeriod=30 Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.371407 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0700b-account-delete-snfrl" event={"ID":"8a9b6914-42e5-4a9c-a4da-de3cee2587e2","Type":"ContainerStarted","Data":"94a096874bc11dec2f4b20e32ad131b600b66e413d85f688b424a26f7fdf23c8"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.374378 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement02ea-account-delete-zbwhh" podStartSLOduration=4.374359978 podStartE2EDuration="4.374359978s" podCreationTimestamp="2025-12-03 17:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:23:07.37146552 +0000 UTC m=+1642.572842381" watchObservedRunningTime="2025-12-03 17:23:07.374359978 +0000 UTC m=+1642.575736829" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.412871 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81" exitCode=0 Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.412913 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025" exitCode=0 Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.413017 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.413067 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.418503 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-55pgs_2923bcc5-72ec-493c-ad10-9f171801ae4a/openstack-network-exporter/0.log" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.418733 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-55pgs" event={"ID":"2923bcc5-72ec-493c-ad10-9f171801ae4a","Type":"ContainerDied","Data":"a094f01db20cff1eb297edaa1f2f22fb6e131a443fa835e281e32bedc079b363"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.418754 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-55pgs" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.418819 4758 scope.go:117] "RemoveContainer" containerID="5877538da9e694dfac7b157eb99747894a3276ae973e31c6d9517ea26fd87fd7" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.422097 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a370910-1957-4bd8-891b-7f1065c5e286-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.423515 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pftgw\" (UniqueName: \"kubernetes.io/projected/d870c33d-67ba-4991-8624-4d0a28cb8542-kube-api-access-pftgw\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.433864 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7a370910-1957-4bd8-891b-7f1065c5e286/ovsdbserver-sb/0.log" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.433964 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7a370910-1957-4bd8-891b-7f1065c5e286","Type":"ContainerDied","Data":"67b7edf337a173465cbd23ee0f7c74958897500d7f7dff5d3ac774d3025af670"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.434615 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.466222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron988a-account-delete-t55hx" event={"ID":"bde9ec06-c65f-45a5-9e90-5a12397f010e","Type":"ContainerStarted","Data":"89739162cf262cead564f1509b480ac13e65ee0c4239567ee0aee5f6c98a8e72"} Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.508053 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.508343 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.522324 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.522788 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.525203 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.525271 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.525462 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d870c33d-67ba-4991-8624-4d0a28cb8542/ovsdbserver-nb/0.log" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.525621 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d870c33d-67ba-4991-8624-4d0a28cb8542","Type":"ContainerDied","Data":"d1d2ed1558de329762a10a5d07018fcffb9a01af10ac51b61adc356e247adcc2"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.525810 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.527946 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.528052 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.533797 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.563790 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.570585 4758 scope.go:117] "RemoveContainer" containerID="da27f6235d4d7b3e9a16f586db08cffa3c4babfe62062989169b8bcb58e3073a" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.570792 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.614266 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" event={"ID":"091fea73-c280-4583-810c-1cfe5f1f0fe1","Type":"ContainerDied","Data":"c023e3687670c040ecfddf0757e36f9bfa704e992f1bb359194cdaf3addb4a89"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.614621 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-s582w" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.637708 4758 generic.go:334] "Generic (PLEG): container finished" podID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerID="a8202adddc1c7cf7ab92d5ed5e61b0495c0d800596e0b0472d95b872e5eca55f" exitCode=0 Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.637830 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"deb8eeb7-c8f9-4324-a518-b837c3ad043c","Type":"ContainerDied","Data":"a8202adddc1c7cf7ab92d5ed5e61b0495c0d800596e0b0472d95b872e5eca55f"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.643504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder14bb-account-delete-ctzv7" event={"ID":"3b4e949a-2e3c-49ff-ba65-f925a9a95336","Type":"ContainerStarted","Data":"1b51b2d524f2fb493fc65f2c72a42dc094eceed8accbcf53dc8de8bdacd6b292"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.645668 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.646038 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicane6b7-account-delete-l2xbs" event={"ID":"cb717c57-e263-449f-896f-38707d61e72c","Type":"ContainerStarted","Data":"92023f680cb8eb12a4f0a2e4e3df3fe6c7304474009d9c11a7ac0023bc23b056"} Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.765519 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-config" (OuterVolumeSpecName: "config") pod "091fea73-c280-4583-810c-1cfe5f1f0fe1" (UID: "091fea73-c280-4583-810c-1cfe5f1f0fe1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.812534 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2923bcc5-72ec-493c-ad10-9f171801ae4a" (UID: "2923bcc5-72ec-493c-ad10-9f171801ae4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.814890 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic90a-account-delete-gg2x2"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.845884 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.847765 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.847808 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.851853 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.853885 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancee859-account-delete-bb4lj"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.861393 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.864173 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.870281 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.870366 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="ebee961b-48e5-4983-bb45-c14bbcc164f8" containerName="nova-cell0-conductor-conductor" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.941392 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "790de883-aaa9-4faa-b18c-61ef4d093ab4" (UID: "790de883-aaa9-4faa-b18c-61ef4d093ab4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.951424 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.956713 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.956758 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: I1203 17:23:07.956769 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.960239 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.974151 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 17:23:07 crc kubenswrapper[4758]: E1203 17:23:07.974259 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.009916 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "091fea73-c280-4583-810c-1cfe5f1f0fe1" (UID: "091fea73-c280-4583-810c-1cfe5f1f0fe1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.019737 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "091fea73-c280-4583-810c-1cfe5f1f0fe1" (UID: "091fea73-c280-4583-810c-1cfe5f1f0fe1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.022025 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "790de883-aaa9-4faa-b18c-61ef4d093ab4" (UID: "790de883-aaa9-4faa-b18c-61ef4d093ab4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.048024 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "091fea73-c280-4583-810c-1cfe5f1f0fe1" (UID: "091fea73-c280-4583-810c-1cfe5f1f0fe1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.061107 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.065258 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.065275 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.065288 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.084828 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "2923bcc5-72ec-493c-ad10-9f171801ae4a" (UID: "2923bcc5-72ec-493c-ad10-9f171801ae4a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.111515 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "091fea73-c280-4583-810c-1cfe5f1f0fe1" (UID: "091fea73-c280-4583-810c-1cfe5f1f0fe1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.137423 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.164347 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "d870c33d-67ba-4991-8624-4d0a28cb8542" (UID: "d870c33d-67ba-4991-8624-4d0a28cb8542"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.168944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "790de883-aaa9-4faa-b18c-61ef4d093ab4" (UID: "790de883-aaa9-4faa-b18c-61ef4d093ab4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.171169 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.171212 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2923bcc5-72ec-493c-ad10-9f171801ae4a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.171222 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/790de883-aaa9-4faa-b18c-61ef4d093ab4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.171232 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091fea73-c280-4583-810c-1cfe5f1f0fe1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.171249 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d870c33d-67ba-4991-8624-4d0a28cb8542-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.477877 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.478426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn6p7\" (UniqueName: \"kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7\") pod \"novacell1a6be-account-delete-tcd65\" (UID: \"154d62cd-e6fa-4b15-ab82-7f486ea22eaa\") " pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.478174 4758 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.478996 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:12.478966126 +0000 UTC m=+1647.680342987 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : configmap "openstack-cell1-scripts" not found Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.494404 4758 projected.go:194] Error preparing data for projected volume kube-api-access-mn6p7 for pod openstack/novacell1a6be-account-delete-tcd65: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.494497 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7 podName:154d62cd-e6fa-4b15-ab82-7f486ea22eaa nodeName:}" failed. No retries permitted until 2025-12-03 17:23:12.494469791 +0000 UTC m=+1647.695846652 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-mn6p7" (UniqueName: "kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7") pod "novacell1a6be-account-delete-tcd65" (UID: "154d62cd-e6fa-4b15-ab82-7f486ea22eaa") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.625900 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41 is running failed: container process not found" containerID="0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.628086 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41 is running failed: container process not found" containerID="0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.631708 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41 is running failed: container process not found" containerID="0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:23:08 crc kubenswrapper[4758]: E1203 17:23:08.631758 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" containerName="nova-scheduler-scheduler" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.699573 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.702439 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee859-account-delete-bb4lj" event={"ID":"5a11fd25-44dc-4ed6-85a5-7f10fc42302a","Type":"ContainerStarted","Data":"210592bc487c79d65a3850b60a1230bf68b4872ed384004affb1f3ec3ce2986d"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.725288 4758 generic.go:334] "Generic (PLEG): container finished" podID="1f328daf-ac93-432b-8645-14b7030a24d1" containerID="e9526364cd771e1d73395320b28267b014ffd81941f74ef5e036bbd181871def" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.725495 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d78d4d4d-rhzqc" event={"ID":"1f328daf-ac93-432b-8645-14b7030a24d1","Type":"ContainerDied","Data":"e9526364cd771e1d73395320b28267b014ffd81941f74ef5e036bbd181871def"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.725617 4758 scope.go:117] "RemoveContainer" containerID="a6273aa4433cec3d63eff07a181e001dfe930ba9327ebff026ed75fb8e277098" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.731628 4758 generic.go:334] "Generic (PLEG): container finished" podID="014153c0-72a5-46ae-9065-ff34c290745a" containerID="484b8a0e84e8e29e4a83fcb1f26cd7fc46d4f41e4674efa1174c310f89876926" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.732816 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement02ea-account-delete-zbwhh" event={"ID":"014153c0-72a5-46ae-9065-ff34c290745a","Type":"ContainerDied","Data":"484b8a0e84e8e29e4a83fcb1f26cd7fc46d4f41e4674efa1174c310f89876926"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.741996 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic90a-account-delete-gg2x2" event={"ID":"fef3d86b-78d4-4b18-9a05-1c2f435f03ed","Type":"ContainerStarted","Data":"79e365a4aa65e515cdb7c413f0cf1bb09a76e0051597affff3892ecd7eebd09b"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.746235 4758 generic.go:334] "Generic (PLEG): container finished" podID="bde9ec06-c65f-45a5-9e90-5a12397f010e" containerID="0bd8d0e46e216ea4fb96179a0303560e480f37a479d27051d4e0b9648d0aaa54" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.746456 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron988a-account-delete-t55hx" event={"ID":"bde9ec06-c65f-45a5-9e90-5a12397f010e","Type":"ContainerDied","Data":"0bd8d0e46e216ea4fb96179a0303560e480f37a479d27051d4e0b9648d0aaa54"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.758530 4758 generic.go:334] "Generic (PLEG): container finished" podID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerID="6b35ae8e6d0ff4c3cd07edce5d23b46583d8b4af2ecf93f7e9477313990aa41c" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.758569 4758 generic.go:334] "Generic (PLEG): container finished" podID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerID="033147cfa4974a6e9865e1dce4a3ab80379ee25004c515e9a54cca5bc2a76e0e" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.758628 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" event={"ID":"583229ad-ae2e-41b5-aad3-3c196b829f40","Type":"ContainerDied","Data":"6b35ae8e6d0ff4c3cd07edce5d23b46583d8b4af2ecf93f7e9477313990aa41c"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.758667 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" event={"ID":"583229ad-ae2e-41b5-aad3-3c196b829f40","Type":"ContainerDied","Data":"033147cfa4974a6e9865e1dce4a3ab80379ee25004c515e9a54cca5bc2a76e0e"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.769716 4758 generic.go:334] "Generic (PLEG): container finished" podID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerID="d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.769939 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"80b1161a-3f70-4956-b13a-1ef693aa2e7c","Type":"ContainerDied","Data":"d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.770240 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"80b1161a-3f70-4956-b13a-1ef693aa2e7c","Type":"ContainerDied","Data":"d6b24e0378e579d80ec113e94b696b20975fa5f080e73db1f5b7c7a8a17eee50"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.770307 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6b24e0378e579d80ec113e94b696b20975fa5f080e73db1f5b7c7a8a17eee50" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.786962 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": dial tcp 10.217.0.205:8775: connect: connection refused" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.787629 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": dial tcp 10.217.0.205:8775: connect: connection refused" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.792696 4758 generic.go:334] "Generic (PLEG): container finished" podID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" containerID="0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.792905 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04b3f0d2-7b01-48a6-b2be-9e205027d1a2","Type":"ContainerDied","Data":"0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.796865 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6df131e1-f13f-4943-95a0-18086525f883","Type":"ContainerDied","Data":"5338ad0f74275da643c904a93cbd3a21618838542f50e1ca0d45757f40ac759e"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.797132 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5338ad0f74275da643c904a93cbd3a21618838542f50e1ca0d45757f40ac759e" Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.801344 4758 generic.go:334] "Generic (PLEG): container finished" podID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerID="ef163a1a1695789c803e3b2391ac85959aee35c4dc735cf405bb6587fe19ed5f" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.801595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b","Type":"ContainerDied","Data":"ef163a1a1695789c803e3b2391ac85959aee35c4dc735cf405bb6587fe19ed5f"} Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.809098 4758 generic.go:334] "Generic (PLEG): container finished" podID="ebee961b-48e5-4983-bb45-c14bbcc164f8" containerID="b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75" exitCode=0 Dec 03 17:23:08 crc kubenswrapper[4758]: I1203 17:23:08.809402 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebee961b-48e5-4983-bb45-c14bbcc164f8","Type":"ContainerDied","Data":"b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.148241 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="790de883-aaa9-4faa-b18c-61ef4d093ab4" path="/var/lib/kubelet/pods/790de883-aaa9-4faa-b18c-61ef4d093ab4/volumes" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.149237 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" path="/var/lib/kubelet/pods/7a370910-1957-4bd8-891b-7f1065c5e286/volumes" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.297385 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.306227 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-s582w"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.365208 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-s582w"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.375600 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.376581 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.376782 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.378743 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-55pgs"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.390909 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.408107 4758 scope.go:117] "RemoveContainer" containerID="ec3db493149a77626469cdaa439ac0d0670962001d5407b77c4d6b653c833931" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.416575 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.427989 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-55pgs"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.430167 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-combined-ca-bundle\") pod \"6df131e1-f13f-4943-95a0-18086525f883\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.430299 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-vencrypt-tls-certs\") pod \"6df131e1-f13f-4943-95a0-18086525f883\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.430375 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97dq7\" (UniqueName: \"kubernetes.io/projected/6df131e1-f13f-4943-95a0-18086525f883-kube-api-access-97dq7\") pod \"6df131e1-f13f-4943-95a0-18086525f883\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.430492 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-nova-novncproxy-tls-certs\") pod \"6df131e1-f13f-4943-95a0-18086525f883\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.430582 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-config-data\") pod \"6df131e1-f13f-4943-95a0-18086525f883\" (UID: \"6df131e1-f13f-4943-95a0-18086525f883\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.456388 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df131e1-f13f-4943-95a0-18086525f883-kube-api-access-97dq7" (OuterVolumeSpecName: "kube-api-access-97dq7") pod "6df131e1-f13f-4943-95a0-18086525f883" (UID: "6df131e1-f13f-4943-95a0-18086525f883"). InnerVolumeSpecName "kube-api-access-97dq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.468981 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.481229 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.494293 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.532950 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggl67\" (UniqueName: \"kubernetes.io/projected/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kube-api-access-ggl67\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533024 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533054 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kolla-config\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533105 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-logs\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533123 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-internal-tls-certs\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533165 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-config-data\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533196 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p69sh\" (UniqueName: \"kubernetes.io/projected/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-kube-api-access-p69sh\") pod \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533215 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-public-tls-certs\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533232 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srjfp\" (UniqueName: \"kubernetes.io/projected/1f328daf-ac93-432b-8645-14b7030a24d1-kube-api-access-srjfp\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533250 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533292 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-operator-scripts\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533326 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-config-data\") pod \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-generated\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533376 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sw8d\" (UniqueName: \"kubernetes.io/projected/ebee961b-48e5-4983-bb45-c14bbcc164f8-kube-api-access-8sw8d\") pod \"ebee961b-48e5-4983-bb45-c14bbcc164f8\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533419 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-combined-ca-bundle\") pod \"ebee961b-48e5-4983-bb45-c14bbcc164f8\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533455 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-config-data\") pod \"ebee961b-48e5-4983-bb45-c14bbcc164f8\" (UID: \"ebee961b-48e5-4983-bb45-c14bbcc164f8\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533478 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-default\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533511 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-scripts\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533559 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data-custom\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533580 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-combined-ca-bundle\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533620 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-internal-tls-certs\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533702 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-galera-tls-certs\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533761 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-etc-machine-id\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533800 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-public-tls-certs\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533827 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-combined-ca-bundle\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533856 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7plrd\" (UniqueName: \"kubernetes.io/projected/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-kube-api-access-7plrd\") pod \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\" (UID: \"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533882 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-combined-ca-bundle\") pod \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\" (UID: \"04b3f0d2-7b01-48a6-b2be-9e205027d1a2\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533936 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-combined-ca-bundle\") pod \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\" (UID: \"80b1161a-3f70-4956-b13a-1ef693aa2e7c\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533959 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f328daf-ac93-432b-8645-14b7030a24d1-logs\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.533986 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-scripts\") pod \"1f328daf-ac93-432b-8645-14b7030a24d1\" (UID: \"1f328daf-ac93-432b-8645-14b7030a24d1\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.534572 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97dq7\" (UniqueName: \"kubernetes.io/projected/6df131e1-f13f-4943-95a0-18086525f883-kube-api-access-97dq7\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.545587 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-logs" (OuterVolumeSpecName: "logs") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.546959 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.547857 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.554325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.565468 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.573579 4758 scope.go:117] "RemoveContainer" containerID="66ede82dc9e0fb893690735dd61d3b58c7e3794bb7c61e6dce24226c9e1b2e55" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.577618 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-kube-api-access-p69sh" (OuterVolumeSpecName: "kube-api-access-p69sh") pod "04b3f0d2-7b01-48a6-b2be-9e205027d1a2" (UID: "04b3f0d2-7b01-48a6-b2be-9e205027d1a2"). InnerVolumeSpecName "kube-api-access-p69sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.578291 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f328daf-ac93-432b-8645-14b7030a24d1-logs" (OuterVolumeSpecName: "logs") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.585974 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-69dc74dcbf-wzmb5" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.164:9696/\": dial tcp 10.217.0.164:9696: connect: connection refused" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.590027 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.603639 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-scripts" (OuterVolumeSpecName: "scripts") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.607848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.618733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebee961b-48e5-4983-bb45-c14bbcc164f8-kube-api-access-8sw8d" (OuterVolumeSpecName: "kube-api-access-8sw8d") pod "ebee961b-48e5-4983-bb45-c14bbcc164f8" (UID: "ebee961b-48e5-4983-bb45-c14bbcc164f8"). InnerVolumeSpecName "kube-api-access-8sw8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.628268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kube-api-access-ggl67" (OuterVolumeSpecName: "kube-api-access-ggl67") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "kube-api-access-ggl67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.640188 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-internal-tls-certs\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.640311 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-etc-swift\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.640525 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-combined-ca-bundle\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.640659 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlbv8\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-kube-api-access-qlbv8\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.640792 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-config-data\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.640843 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-log-httpd\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.641037 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-run-httpd\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.641068 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-public-tls-certs\") pod \"583229ad-ae2e-41b5-aad3-3c196b829f40\" (UID: \"583229ad-ae2e-41b5-aad3-3c196b829f40\") " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642127 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642160 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642174 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sw8d\" (UniqueName: \"kubernetes.io/projected/ebee961b-48e5-4983-bb45-c14bbcc164f8-kube-api-access-8sw8d\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642185 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642195 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642206 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642217 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642227 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f328daf-ac93-432b-8645-14b7030a24d1-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642238 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggl67\" (UniqueName: \"kubernetes.io/projected/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kube-api-access-ggl67\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642248 4758 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/80b1161a-3f70-4956-b13a-1ef693aa2e7c-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642259 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642269 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p69sh\" (UniqueName: \"kubernetes.io/projected/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-kube-api-access-p69sh\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.642889 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.643317 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.682692 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f328daf-ac93-432b-8645-14b7030a24d1-kube-api-access-srjfp" (OuterVolumeSpecName: "kube-api-access-srjfp") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "kube-api-access-srjfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.698459 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-scripts" (OuterVolumeSpecName: "scripts") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.745141 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-kube-api-access-7plrd" (OuterVolumeSpecName: "kube-api-access-7plrd") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "kube-api-access-7plrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.747009 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.747063 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.747076 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srjfp\" (UniqueName: \"kubernetes.io/projected/1f328daf-ac93-432b-8645-14b7030a24d1-kube-api-access-srjfp\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.747086 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/583229ad-ae2e-41b5-aad3-3c196b829f40-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.774021 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-kube-api-access-qlbv8" (OuterVolumeSpecName: "kube-api-access-qlbv8") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "kube-api-access-qlbv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.782428 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.835485 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.840024 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d78d4d4d-rhzqc" event={"ID":"1f328daf-ac93-432b-8645-14b7030a24d1","Type":"ContainerDied","Data":"d188692e4f7d254f836db7bf4c350e95d6c6d3c30b3458283cab4b30afa448da"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.840182 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68d78d4d4d-rhzqc" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.847120 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0700b-account-delete-snfrl" event={"ID":"8a9b6914-42e5-4a9c-a4da-de3cee2587e2","Type":"ContainerStarted","Data":"0cb9a87be8ad59f9ff8dd09d281da1a8164d0ffdaca0aabbf9f6081178c0baf1"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.854706 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7plrd\" (UniqueName: \"kubernetes.io/projected/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-kube-api-access-7plrd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.854971 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlbv8\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-kube-api-access-qlbv8\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.855006 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.855039 4758 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/583229ad-ae2e-41b5-aad3-3c196b829f40-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.871547 4758 generic.go:334] "Generic (PLEG): container finished" podID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerID="b226eaf6e02f5bf23bb058d2357c7eb31da688995dc12dbdc9fe34fe8dbc7091" exitCode=0 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.871808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b49b1d3-cbb4-49da-8244-69e15595da33","Type":"ContainerDied","Data":"b226eaf6e02f5bf23bb058d2357c7eb31da688995dc12dbdc9fe34fe8dbc7091"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.873972 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.874945 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="sg-core" containerID="cri-o://a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad" gracePeriod=30 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.875248 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="proxy-httpd" containerID="cri-o://922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a" gracePeriod=30 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.878058 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-notification-agent" containerID="cri-o://d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c" gracePeriod=30 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.879425 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-central-agent" containerID="cri-o://e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388" gracePeriod=30 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.904794 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d50c7a0d-80f6-4c8c-a516-29dbe3a6397b","Type":"ContainerDied","Data":"76de1c5f74c1f80efeadc927dff83e66fb87f4001807e5daeeafe7f4f580cfc3"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.904849 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.906830 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.916984 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="82a1c0c2-3a27-43cc-8d41-d3c7508ab441" containerName="kube-state-metrics" containerID="cri-o://f6091694aad9cdd8c04fa3979ee4bf04648cfc927b2b56cd01f71c7f65050343" gracePeriod=30 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.925610 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell0700b-account-delete-snfrl" podStartSLOduration=6.925587138 podStartE2EDuration="6.925587138s" podCreationTimestamp="2025-12-03 17:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:23:09.882098793 +0000 UTC m=+1645.083475654" watchObservedRunningTime="2025-12-03 17:23:09.925587138 +0000 UTC m=+1645.126963999" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.955329 4758 generic.go:334] "Generic (PLEG): container finished" podID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerID="9d2c93db5d7424078ddb1bb905ea651dea2607f1f94df3e14e8eaf9f6eec920f" exitCode=0 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.955738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be629420-0e78-44d4-9ca1-4ee1601d4ca6","Type":"ContainerDied","Data":"9d2c93db5d7424078ddb1bb905ea651dea2607f1f94df3e14e8eaf9f6eec920f"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.972994 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.974049 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebee961b-48e5-4983-bb45-c14bbcc164f8","Type":"ContainerDied","Data":"8f73a844ae769c58acc396ebbb266655e2232b126e23cdb30ff2999baaaabeca"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.993324 4758 generic.go:334] "Generic (PLEG): container finished" podID="84469de1-c706-4442-8610-31188889dad1" containerID="08867454fb2022e6047c48092dc940ce2f60d0e3bb28b9647d08e3cc414e048e" exitCode=0 Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.993728 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84469de1-c706-4442-8610-31188889dad1","Type":"ContainerDied","Data":"08867454fb2022e6047c48092dc940ce2f60d0e3bb28b9647d08e3cc414e048e"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.993887 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84469de1-c706-4442-8610-31188889dad1","Type":"ContainerDied","Data":"16e24156cb150c6e3db9cd76a8a264ed1454fe3279b2b439474c2e9c18185aec"} Dec 03 17:23:09 crc kubenswrapper[4758]: I1203 17:23:09.993983 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16e24156cb150c6e3db9cd76a8a264ed1454fe3279b2b439474c2e9c18185aec" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.006380 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04b3f0d2-7b01-48a6-b2be-9e205027d1a2","Type":"ContainerDied","Data":"ed8709865c5ffef4e83ed1e2430a6460910b2286370ca8d329cc488dfaf14eff"} Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.006587 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.048266 4758 generic.go:334] "Generic (PLEG): container finished" podID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerID="d0fc34ea02fb5712f5c26717176864428bc2e389b2d8605da62e87ecf5f8f023" exitCode=0 Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.048567 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b461991e-58f6-4ff3-81d3-9cf2288d9b65","Type":"ContainerDied","Data":"d0fc34ea02fb5712f5c26717176864428bc2e389b2d8605da62e87ecf5f8f023"} Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.050843 4758 generic.go:334] "Generic (PLEG): container finished" podID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerID="c1e072acd4c62b59d1cb154ec06ad176f5ed4d1bdd3cc652119c966d7118b7f9" exitCode=0 Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.050947 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-576b5fdbd4-qmczw" event={"ID":"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665","Type":"ContainerDied","Data":"c1e072acd4c62b59d1cb154ec06ad176f5ed4d1bdd3cc652119c966d7118b7f9"} Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.051306 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.051607 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="0754ea69-5007-4e51-bd22-0efd13954199" containerName="memcached" containerID="cri-o://2071a07b9d8bd2fffbcfc63e0cd6e8e249aa95143b551ab078f950658b27507b" gracePeriod=30 Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.062119 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.177:9292/healthcheck\": dial tcp 10.217.0.177:9292: connect: connection refused" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.062892 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.177:9292/healthcheck\": dial tcp 10.217.0.177:9292: connect: connection refused" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.063016 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" event={"ID":"583229ad-ae2e-41b5-aad3-3c196b829f40","Type":"ContainerDied","Data":"944f9d6d0e211ceb4b3ce57796bfff3ea944d6469168a780a09799cfa3845c76"} Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.063228 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6fc8f59f6f-7zxl4" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.088917 4758 generic.go:334] "Generic (PLEG): container finished" podID="3b4e949a-2e3c-49ff-ba65-f925a9a95336" containerID="421b471351888521cc265529cdcf5204bbfd978298865928a87d0b8d3f8072bd" exitCode=0 Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.089051 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1a6be-account-delete-tcd65" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.090083 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder14bb-account-delete-ctzv7" event={"ID":"3b4e949a-2e3c-49ff-ba65-f925a9a95336","Type":"ContainerDied","Data":"421b471351888521cc265529cdcf5204bbfd978298865928a87d0b8d3f8072bd"} Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.090373 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.090538 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.110243 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2rcnb"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.315629 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-stqc7"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.325157 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-config-data" (OuterVolumeSpecName: "config-data") pod "6df131e1-f13f-4943-95a0-18086525f883" (UID: "6df131e1-f13f-4943-95a0-18086525f883"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.348844 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-stqc7"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.368478 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2rcnb"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.386057 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-584ffbf4c6-g7bkv"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.389919 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-584ffbf4c6-g7bkv" podUID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" containerName="keystone-api" containerID="cri-o://1d73cb8d75674ff7493a3d6896b8923851258760396fa1c116f8e006497ff57c" gracePeriod=30 Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.425133 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.428962 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.432498 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04b3f0d2-7b01-48a6-b2be-9e205027d1a2" (UID: "04b3f0d2-7b01-48a6-b2be-9e205027d1a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.456720 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-cpjwp"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.464707 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-cpjwp"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.494265 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data" (OuterVolumeSpecName: "config-data") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.494259 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.507928 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8aca-account-create-update-d8h7n"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.522283 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8aca-account-create-update-d8h7n"] Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.532491 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.534335 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.534357 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.534369 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.534380 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.662170 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.738706 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.759296 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-config-data" (OuterVolumeSpecName: "config-data") pod "ebee961b-48e5-4983-bb45-c14bbcc164f8" (UID: "ebee961b-48e5-4983-bb45-c14bbcc164f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.764009 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebee961b-48e5-4983-bb45-c14bbcc164f8" (UID: "ebee961b-48e5-4983-bb45-c14bbcc164f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.770767 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.844388 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.844425 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebee961b-48e5-4983-bb45-c14bbcc164f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.844435 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: E1203 17:23:10.844574 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 17:23:10 crc kubenswrapper[4758]: E1203 17:23:10.844817 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data podName:070406cc-6d65-4fd0-81d7-264adead6bbd nodeName:}" failed. No retries permitted until 2025-12-03 17:23:18.844788957 +0000 UTC m=+1654.046165818 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data") pod "rabbitmq-server-0" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd") : configmap "rabbitmq-config-data" not found Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.858893 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-config-data" (OuterVolumeSpecName: "config-data") pod "04b3f0d2-7b01-48a6-b2be-9e205027d1a2" (UID: "04b3f0d2-7b01-48a6-b2be-9e205027d1a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.862979 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "6df131e1-f13f-4943-95a0-18086525f883" (UID: "6df131e1-f13f-4943-95a0-18086525f883"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.872273 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.898970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6df131e1-f13f-4943-95a0-18086525f883" (UID: "6df131e1-f13f-4943-95a0-18086525f883"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.951724 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.951782 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b3f0d2-7b01-48a6-b2be-9e205027d1a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.951804 4758 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.951834 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:10 crc kubenswrapper[4758]: E1203 17:23:10.951989 4758 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 03 17:23:10 crc kubenswrapper[4758]: E1203 17:23:10.952070 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts podName:aac23ac3-72ea-49a7-9ded-6fabe28ff690 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:18.952049112 +0000 UTC m=+1654.153425973 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts") pod "ovn-controller-5czb6" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690") : configmap "ovncontroller-scripts" not found Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.954463 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" (UID: "d50c7a0d-80f6-4c8c-a516-29dbe3a6397b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.965943 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="galera" containerID="cri-o://d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea" gracePeriod=30 Dec 03 17:23:10 crc kubenswrapper[4758]: I1203 17:23:10.970012 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "80b1161a-3f70-4956-b13a-1ef693aa2e7c" (UID: "80b1161a-3f70-4956-b13a-1ef693aa2e7c"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.056057 4758 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b1161a-3f70-4956-b13a-1ef693aa2e7c-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.056089 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.070875 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "6df131e1-f13f-4943-95a0-18086525f883" (UID: "6df131e1-f13f-4943-95a0-18086525f883"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.071889 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.104666 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement02ea-account-delete-zbwhh" event={"ID":"014153c0-72a5-46ae-9065-ff34c290745a","Type":"ContainerDied","Data":"020331e515ccd2014f34f68d90c60679466b1826ddc48a69b7e2e4da84f2079e"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.105533 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="020331e515ccd2014f34f68d90c60679466b1826ddc48a69b7e2e4da84f2079e" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.116013 4758 generic.go:334] "Generic (PLEG): container finished" podID="7bb44365-2dda-4769-813a-712f362fff6b" containerID="922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a" exitCode=0 Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.116053 4758 generic.go:334] "Generic (PLEG): container finished" podID="7bb44365-2dda-4769-813a-712f362fff6b" containerID="a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad" exitCode=2 Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.116064 4758 generic.go:334] "Generic (PLEG): container finished" podID="7bb44365-2dda-4769-813a-712f362fff6b" containerID="e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388" exitCode=0 Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.139059 4758 generic.go:334] "Generic (PLEG): container finished" podID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerID="df8bc0faf1cabc23c62d9aea7423ca778a9372ef7f9e311140e90eae1b53b014" exitCode=0 Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.140251 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" path="/var/lib/kubelet/pods/091fea73-c280-4583-810c-1cfe5f1f0fe1/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.141055 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2923bcc5-72ec-493c-ad10-9f171801ae4a" path="/var/lib/kubelet/pods/2923bcc5-72ec-493c-ad10-9f171801ae4a/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.142040 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84012fbf-daba-42b5-9c41-cd5a5a91da17" path="/var/lib/kubelet/pods/84012fbf-daba-42b5-9c41-cd5a5a91da17/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.143422 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e2e52f9-0802-4065-9630-2a2200060db4" path="/var/lib/kubelet/pods/8e2e52f9-0802-4065-9630-2a2200060db4/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.149849 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glancee859-account-delete-bb4lj" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.150536 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b902e850-9e45-4d1d-bc6e-0d71d5830db9" path="/var/lib/kubelet/pods/b902e850-9e45-4d1d-bc6e-0d71d5830db9/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.152374 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" path="/var/lib/kubelet/pods/d870c33d-67ba-4991-8624-4d0a28cb8542/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.153848 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fce262ba-d313-4515-ad6b-e7137a980109" path="/var/lib/kubelet/pods/fce262ba-d313-4515-ad6b-e7137a980109/volumes" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.157354 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicane6b7-account-delete-l2xbs" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.160042 4758 generic.go:334] "Generic (PLEG): container finished" podID="82a1c0c2-3a27-43cc-8d41-d3c7508ab441" containerID="f6091694aad9cdd8c04fa3979ee4bf04648cfc927b2b56cd01f71c7f65050343" exitCode=2 Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.170507 4758 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df131e1-f13f-4943-95a0-18086525f883-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.170540 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.161747 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.171847 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-config-data" (OuterVolumeSpecName: "config-data") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.198441 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-config-data" (OuterVolumeSpecName: "config-data") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.202247 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapic90a-account-delete-gg2x2" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.208823 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glancee859-account-delete-bb4lj" podStartSLOduration=8.208787931 podStartE2EDuration="8.208787931s" podCreationTimestamp="2025-12-03 17:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:23:11.165839431 +0000 UTC m=+1646.367216302" watchObservedRunningTime="2025-12-03 17:23:11.208787931 +0000 UTC m=+1646.410164812" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.214165 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell0700b-account-delete-snfrl" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.225725 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbicane6b7-account-delete-l2xbs" podStartSLOduration=8.225673993 podStartE2EDuration="8.225673993s" podCreationTimestamp="2025-12-03 17:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:23:11.190371787 +0000 UTC m=+1646.391748648" watchObservedRunningTime="2025-12-03 17:23:11.225673993 +0000 UTC m=+1646.427050854" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.228495 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "583229ad-ae2e-41b5-aad3-3c196b829f40" (UID: "583229ad-ae2e-41b5-aad3-3c196b829f40"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.241359 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapic90a-account-delete-gg2x2" podStartSLOduration=8.241326713 podStartE2EDuration="8.241326713s" podCreationTimestamp="2025-12-03 17:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:23:11.209269194 +0000 UTC m=+1646.410646055" watchObservedRunningTime="2025-12-03 17:23:11.241326713 +0000 UTC m=+1646.442703574" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273470 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerDied","Data":"922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273536 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerDied","Data":"a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273569 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerDied","Data":"e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273581 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"deb8eeb7-c8f9-4324-a518-b837c3ad043c","Type":"ContainerDied","Data":"df8bc0faf1cabc23c62d9aea7423ca778a9372ef7f9e311140e90eae1b53b014"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273597 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-576b5fdbd4-qmczw" event={"ID":"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665","Type":"ContainerDied","Data":"1a30804cb4612cabd8244c764409e4ecfb725677392161471dcd1712c9a39290"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273614 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a30804cb4612cabd8244c764409e4ecfb725677392161471dcd1712c9a39290" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b49b1d3-cbb4-49da-8244-69e15595da33","Type":"ContainerDied","Data":"42229c66216558885e65ed8defa507e5e594d94a41d371a6cf2e4e46a13a6971"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273654 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42229c66216558885e65ed8defa507e5e594d94a41d371a6cf2e4e46a13a6971" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.273664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee859-account-delete-bb4lj" event={"ID":"5a11fd25-44dc-4ed6-85a5-7f10fc42302a","Type":"ContainerStarted","Data":"d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.275564 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicane6b7-account-delete-l2xbs" event={"ID":"cb717c57-e263-449f-896f-38707d61e72c","Type":"ContainerStarted","Data":"3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.275640 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82a1c0c2-3a27-43cc-8d41-d3c7508ab441","Type":"ContainerDied","Data":"f6091694aad9cdd8c04fa3979ee4bf04648cfc927b2b56cd01f71c7f65050343"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.275664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic90a-account-delete-gg2x2" event={"ID":"fef3d86b-78d4-4b18-9a05-1c2f435f03ed","Type":"ContainerStarted","Data":"6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.275706 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b461991e-58f6-4ff3-81d3-9cf2288d9b65","Type":"ContainerDied","Data":"ecbd0b931eee3f79ade53d3def7b3127f4c981c4b84b65ae184dfb6225502c64"} Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.275724 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecbd0b931eee3f79ade53d3def7b3127f4c981c4b84b65ae184dfb6225502c64" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.277150 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.277215 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.277301 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:11.777271376 +0000 UTC m=+1646.978648417 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.278574 4758 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.278619 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data podName:e518390e-b4c4-49bb-913f-ddfc331456ab nodeName:}" failed. No retries permitted until 2025-12-03 17:23:19.278605592 +0000 UTC m=+1654.479982603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data") pod "rabbitmq-cell1-server-0" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab") : configmap "rabbitmq-cell1-config-data" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.278639 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.279122 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.279157 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.279789 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:23:11.779774863 +0000 UTC m=+1646.981151914 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.279846 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:23:11.779815944 +0000 UTC m=+1646.981192805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.279908 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.279965 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.280911 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:23:11.780854912 +0000 UTC m=+1646.982231773 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.281040 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/583229ad-ae2e-41b5-aad3-3c196b829f40-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.285196 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.301736 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.340722 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1f328daf-ac93-432b-8645-14b7030a24d1" (UID: "1f328daf-ac93-432b-8645-14b7030a24d1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.341137 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.363974 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.364051 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.373230 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.382113 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84469de1-c706-4442-8610-31188889dad1-logs\") pod \"84469de1-c706-4442-8610-31188889dad1\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.382250 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtltd\" (UniqueName: \"kubernetes.io/projected/84469de1-c706-4442-8610-31188889dad1-kube-api-access-xtltd\") pod \"84469de1-c706-4442-8610-31188889dad1\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.385366 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84469de1-c706-4442-8610-31188889dad1-logs" (OuterVolumeSpecName: "logs") pod "84469de1-c706-4442-8610-31188889dad1" (UID: "84469de1-c706-4442-8610-31188889dad1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.392910 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-nova-metadata-tls-certs\") pod \"84469de1-c706-4442-8610-31188889dad1\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.393004 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-config-data\") pod \"84469de1-c706-4442-8610-31188889dad1\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.393090 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-combined-ca-bundle\") pod \"84469de1-c706-4442-8610-31188889dad1\" (UID: \"84469de1-c706-4442-8610-31188889dad1\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.394493 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84469de1-c706-4442-8610-31188889dad1-kube-api-access-xtltd" (OuterVolumeSpecName: "kube-api-access-xtltd") pod "84469de1-c706-4442-8610-31188889dad1" (UID: "84469de1-c706-4442-8610-31188889dad1"). InnerVolumeSpecName "kube-api-access-xtltd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.393000 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.398302 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.398344 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f328daf-ac93-432b-8645-14b7030a24d1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.398359 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84469de1-c706-4442-8610-31188889dad1-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.398372 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtltd\" (UniqueName: \"kubernetes.io/projected/84469de1-c706-4442-8610-31188889dad1-kube-api-access-xtltd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.404090 4758 scope.go:117] "RemoveContainer" containerID="5659d420ac31a290e23ff03a8dee2a674a17edcf17f81a32fdb9e77fc73d7428" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.410619 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1a6be-account-delete-tcd65"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.426122 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.444518 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell1a6be-account-delete-tcd65"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.455010 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.466022 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84469de1-c706-4442-8610-31188889dad1" (UID: "84469de1-c706-4442-8610-31188889dad1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.473569 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.475865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-config-data" (OuterVolumeSpecName: "config-data") pod "84469de1-c706-4442-8610-31188889dad1" (UID: "84469de1-c706-4442-8610-31188889dad1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.478414 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.502985 4758 scope.go:117] "RemoveContainer" containerID="3c175f3c3abbd9f78b887a0d32c0fdf73322a6b77388cdd343be7a3ed9da0d73" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.503506 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b461991e-58f6-4ff3-81d3-9cf2288d9b65-logs\") pod \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.503567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/014153c0-72a5-46ae-9065-ff34c290745a-operator-scripts\") pod \"014153c0-72a5-46ae-9065-ff34c290745a\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.503615 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.503654 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-internal-tls-certs\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.504871 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b461991e-58f6-4ff3-81d3-9cf2288d9b65-logs" (OuterVolumeSpecName: "logs") pod "b461991e-58f6-4ff3-81d3-9cf2288d9b65" (UID: "b461991e-58f6-4ff3-81d3-9cf2288d9b65"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.504918 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/014153c0-72a5-46ae-9065-ff34c290745a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "014153c0-72a5-46ae-9065-ff34c290745a" (UID: "014153c0-72a5-46ae-9065-ff34c290745a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505291 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-public-tls-certs\") pod \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7mbs\" (UniqueName: \"kubernetes.io/projected/014153c0-72a5-46ae-9065-ff34c290745a-kube-api-access-w7mbs\") pod \"014153c0-72a5-46ae-9065-ff34c290745a\" (UID: \"014153c0-72a5-46ae-9065-ff34c290745a\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505475 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-combined-ca-bundle\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505523 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-logs\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505600 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-combined-ca-bundle\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505635 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6swb6\" (UniqueName: \"kubernetes.io/projected/b461991e-58f6-4ff3-81d3-9cf2288d9b65-kube-api-access-6swb6\") pod \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505714 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-scripts\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505751 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-combined-ca-bundle\") pod \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505886 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-internal-tls-certs\") pod \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.505978 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-internal-tls-certs\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506045 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87nfc\" (UniqueName: \"kubernetes.io/projected/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-kube-api-access-87nfc\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506129 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-httpd-run\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506150 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-config-data\") pod \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\" (UID: \"b461991e-58f6-4ff3-81d3-9cf2288d9b65\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506174 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg4gr\" (UniqueName: \"kubernetes.io/projected/1b49b1d3-cbb4-49da-8244-69e15595da33-kube-api-access-mg4gr\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506214 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-public-tls-certs\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506236 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506258 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-logs\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506293 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-config-data\") pod \"1b49b1d3-cbb4-49da-8244-69e15595da33\" (UID: \"1b49b1d3-cbb4-49da-8244-69e15595da33\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.506318 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data-custom\") pod \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\" (UID: \"d0cc9a7c-fe92-4cfe-b8f6-a523db38b665\") " Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.506936 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.507023 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "84469de1-c706-4442-8610-31188889dad1" (UID: "84469de1-c706-4442-8610-31188889dad1"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.507164 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.507194 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.507212 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b461991e-58f6-4ff3-81d3-9cf2288d9b65-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.507224 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/014153c0-72a5-46ae-9065-ff34c290745a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.507423 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.519059 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.519984 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-logs" (OuterVolumeSpecName: "logs") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.521431 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.530244 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.530322 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-logs" (OuterVolumeSpecName: "logs") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.530340 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerName="nova-cell1-conductor-conductor" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.531323 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.555819 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b49b1d3-cbb4-49da-8244-69e15595da33-kube-api-access-mg4gr" (OuterVolumeSpecName: "kube-api-access-mg4gr") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "kube-api-access-mg4gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.557514 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.561572 4758 scope.go:117] "RemoveContainer" containerID="d2e1df737da67a9872c32625a37897f90c909e09e9f108e0525e5a4192296e68" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.573796 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.579391 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.579715 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b461991e-58f6-4ff3-81d3-9cf2288d9b65-kube-api-access-6swb6" (OuterVolumeSpecName: "kube-api-access-6swb6") pod "b461991e-58f6-4ff3-81d3-9cf2288d9b65" (UID: "b461991e-58f6-4ff3-81d3-9cf2288d9b65"). InnerVolumeSpecName "kube-api-access-6swb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.579845 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-kube-api-access-87nfc" (OuterVolumeSpecName: "kube-api-access-87nfc") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "kube-api-access-87nfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.585180 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-scripts" (OuterVolumeSpecName: "scripts") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.586902 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/014153c0-72a5-46ae-9065-ff34c290745a-kube-api-access-w7mbs" (OuterVolumeSpecName: "kube-api-access-w7mbs") pod "014153c0-72a5-46ae-9065-ff34c290745a" (UID: "014153c0-72a5-46ae-9065-ff34c290745a"). InnerVolumeSpecName "kube-api-access-w7mbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.590980 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.620046 4758 scope.go:117] "RemoveContainer" containerID="e9526364cd771e1d73395320b28267b014ffd81941f74ef5e036bbd181871def" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.622503 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.622887 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-public-tls-certs\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623047 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-config\") pod \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623150 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v82zd\" (UniqueName: \"kubernetes.io/projected/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-api-access-v82zd\") pod \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-httpd-run\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-config-data\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623661 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-certs\") pod \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623792 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde9ec06-c65f-45a5-9e90-5a12397f010e-operator-scripts\") pod \"bde9ec06-c65f-45a5-9e90-5a12397f010e\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623882 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-logs\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.623978 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9gs4\" (UniqueName: \"kubernetes.io/projected/be629420-0e78-44d4-9ca1-4ee1601d4ca6-kube-api-access-b9gs4\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.624119 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-scripts\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.624211 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-combined-ca-bundle\") pod \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\" (UID: \"be629420-0e78-44d4-9ca1-4ee1601d4ca6\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.624309 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-combined-ca-bundle\") pod \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\" (UID: \"82a1c0c2-3a27-43cc-8d41-d3c7508ab441\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.624381 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86r88\" (UniqueName: \"kubernetes.io/projected/bde9ec06-c65f-45a5-9e90-5a12397f010e-kube-api-access-86r88\") pod \"bde9ec06-c65f-45a5-9e90-5a12397f010e\" (UID: \"bde9ec06-c65f-45a5-9e90-5a12397f010e\") " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625195 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625273 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg4gr\" (UniqueName: \"kubernetes.io/projected/1b49b1d3-cbb4-49da-8244-69e15595da33-kube-api-access-mg4gr\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625413 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625483 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b49b1d3-cbb4-49da-8244-69e15595da33-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625540 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn6p7\" (UniqueName: \"kubernetes.io/projected/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-kube-api-access-mn6p7\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625596 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625651 4758 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/84469de1-c706-4442-8610-31188889dad1-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625730 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7mbs\" (UniqueName: \"kubernetes.io/projected/014153c0-72a5-46ae-9065-ff34c290745a-kube-api-access-w7mbs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625802 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/154d62cd-e6fa-4b15-ab82-7f486ea22eaa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625857 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.625984 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6swb6\" (UniqueName: \"kubernetes.io/projected/b461991e-58f6-4ff3-81d3-9cf2288d9b65-kube-api-access-6swb6\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.626040 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.626094 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87nfc\" (UniqueName: \"kubernetes.io/projected/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-kube-api-access-87nfc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.626940 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-logs" (OuterVolumeSpecName: "logs") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.627273 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde9ec06-c65f-45a5-9e90-5a12397f010e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bde9ec06-c65f-45a5-9e90-5a12397f010e" (UID: "bde9ec06-c65f-45a5-9e90-5a12397f010e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.627640 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.644619 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.649198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be629420-0e78-44d4-9ca1-4ee1601d4ca6-kube-api-access-b9gs4" (OuterVolumeSpecName: "kube-api-access-b9gs4") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "kube-api-access-b9gs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.660443 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.707206 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-scripts" (OuterVolumeSpecName: "scripts") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.717097 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.721484 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde9ec06-c65f-45a5-9e90-5a12397f010e-kube-api-access-86r88" (OuterVolumeSpecName: "kube-api-access-86r88") pod "bde9ec06-c65f-45a5-9e90-5a12397f010e" (UID: "bde9ec06-c65f-45a5-9e90-5a12397f010e"). InnerVolumeSpecName "kube-api-access-86r88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733466 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733505 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde9ec06-c65f-45a5-9e90-5a12397f010e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733517 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be629420-0e78-44d4-9ca1-4ee1601d4ca6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733527 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9gs4\" (UniqueName: \"kubernetes.io/projected/be629420-0e78-44d4-9ca1-4ee1601d4ca6-kube-api-access-b9gs4\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733536 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733546 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86r88\" (UniqueName: \"kubernetes.io/projected/bde9ec06-c65f-45a5-9e90-5a12397f010e-kube-api-access-86r88\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.733569 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.735274 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.735533 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-api-access-v82zd" (OuterVolumeSpecName: "kube-api-access-v82zd") pod "82a1c0c2-3a27-43cc-8d41-d3c7508ab441" (UID: "82a1c0c2-3a27-43cc-8d41-d3c7508ab441"). InnerVolumeSpecName "kube-api-access-v82zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.781016 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-68d78d4d4d-rhzqc"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.821592 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-68d78d4d4d-rhzqc"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.837288 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.837319 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v82zd\" (UniqueName: \"kubernetes.io/projected/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-api-access-v82zd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.837403 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.837475 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:23:12.837450406 +0000 UTC m=+1648.038827267 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.837944 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.837988 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:23:12.83797607 +0000 UTC m=+1648.039352931 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.838893 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.839023 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:12.838988937 +0000 UTC m=+1648.040365968 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.841925 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: E1203 17:23:11.842042 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:23:12.842015179 +0000 UTC m=+1648.043392040 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.852738 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6fc8f59f6f-7zxl4"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.861994 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-6fc8f59f6f-7zxl4"] Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.865397 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.877851 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b461991e-58f6-4ff3-81d3-9cf2288d9b65" (UID: "b461991e-58f6-4ff3-81d3-9cf2288d9b65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.896412 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.900579 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b461991e-58f6-4ff3-81d3-9cf2288d9b65" (UID: "b461991e-58f6-4ff3-81d3-9cf2288d9b65"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.907912 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "82a1c0c2-3a27-43cc-8d41-d3c7508ab441" (UID: "82a1c0c2-3a27-43cc-8d41-d3c7508ab441"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.914790 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82a1c0c2-3a27-43cc-8d41-d3c7508ab441" (UID: "82a1c0c2-3a27-43cc-8d41-d3c7508ab441"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.916199 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.919538 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-config-data" (OuterVolumeSpecName: "config-data") pod "1b49b1d3-cbb4-49da-8244-69e15595da33" (UID: "1b49b1d3-cbb4-49da-8244-69e15595da33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.924119 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.943275 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-config-data" (OuterVolumeSpecName: "config-data") pod "b461991e-58f6-4ff3-81d3-9cf2288d9b65" (UID: "b461991e-58f6-4ff3-81d3-9cf2288d9b65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945515 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945579 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945597 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945616 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945631 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945645 4758 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945658 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945671 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945704 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.945717 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49b1d3-cbb4-49da-8244-69e15595da33-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.955730 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.958957 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.966442 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "82a1c0c2-3a27-43cc-8d41-d3c7508ab441" (UID: "82a1c0c2-3a27-43cc-8d41-d3c7508ab441"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.968936 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data" (OuterVolumeSpecName: "config-data") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.985835 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" (UID: "d0cc9a7c-fe92-4cfe-b8f6-a523db38b665"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.990867 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b461991e-58f6-4ff3-81d3-9cf2288d9b65" (UID: "b461991e-58f6-4ff3-81d3-9cf2288d9b65"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:11 crc kubenswrapper[4758]: I1203 17:23:11.994173 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.004205 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-config-data" (OuterVolumeSpecName: "config-data") pod "be629420-0e78-44d4-9ca1-4ee1601d4ca6" (UID: "be629420-0e78-44d4-9ca1-4ee1601d4ca6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.047900 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.047945 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.047958 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.047969 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.047979 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.047989 4758 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82a1c0c2-3a27-43cc-8d41-d3c7508ab441-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.048001 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be629420-0e78-44d4-9ca1-4ee1601d4ca6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.048011 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b461991e-58f6-4ff3-81d3-9cf2288d9b65-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.168845 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.171462 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.178941 4758 scope.go:117] "RemoveContainer" containerID="7ff34332c6d8c7344235f023f03dbde628713fc91c197634ffc70e74d0fc3ae7" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.214049 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron988a-account-delete-t55hx" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.214448 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron988a-account-delete-t55hx" event={"ID":"bde9ec06-c65f-45a5-9e90-5a12397f010e","Type":"ContainerDied","Data":"89739162cf262cead564f1509b480ac13e65ee0c4239567ee0aee5f6c98a8e72"} Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.214529 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89739162cf262cead564f1509b480ac13e65ee0c4239567ee0aee5f6c98a8e72" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.219374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"deb8eeb7-c8f9-4324-a518-b837c3ad043c","Type":"ContainerDied","Data":"f525490acf63bb49edc7ef1a5ca45e4acb7b684cf0db3b5fc21956513876cd11"} Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.219742 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.227163 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder14bb-account-delete-ctzv7" event={"ID":"3b4e949a-2e3c-49ff-ba65-f925a9a95336","Type":"ContainerDied","Data":"1b51b2d524f2fb493fc65f2c72a42dc094eceed8accbcf53dc8de8bdacd6b292"} Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.227218 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b51b2d524f2fb493fc65f2c72a42dc094eceed8accbcf53dc8de8bdacd6b292" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.227342 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder14bb-account-delete-ctzv7" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.229910 4758 generic.go:334] "Generic (PLEG): container finished" podID="0754ea69-5007-4e51-bd22-0efd13954199" containerID="2071a07b9d8bd2fffbcfc63e0cd6e8e249aa95143b551ab078f950658b27507b" exitCode=0 Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.229953 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0754ea69-5007-4e51-bd22-0efd13954199","Type":"ContainerDied","Data":"2071a07b9d8bd2fffbcfc63e0cd6e8e249aa95143b551ab078f950658b27507b"} Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.231494 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82a1c0c2-3a27-43cc-8d41-d3c7508ab441","Type":"ContainerDied","Data":"cf1ed8587f64f1208848cb64622fd7fe99899020ec341d6cf2a3359b753fc420"} Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.231555 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.250922 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4e949a-2e3c-49ff-ba65-f925a9a95336-operator-scripts\") pod \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.251183 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hswgp\" (UniqueName: \"kubernetes.io/projected/3b4e949a-2e3c-49ff-ba65-f925a9a95336-kube-api-access-hswgp\") pod \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\" (UID: \"3b4e949a-2e3c-49ff-ba65-f925a9a95336\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.252430 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4e949a-2e3c-49ff-ba65-f925a9a95336-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3b4e949a-2e3c-49ff-ba65-f925a9a95336" (UID: "3b4e949a-2e3c-49ff-ba65-f925a9a95336"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.255052 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be629420-0e78-44d4-9ca1-4ee1601d4ca6","Type":"ContainerDied","Data":"a416b337d477e7ddcf7f04a362cfc7fd6501517331e259aa030296f63e2f24c5"} Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.255165 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.259704 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.261261 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicane6b7-account-delete-l2xbs" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.261804 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement02ea-account-delete-zbwhh" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.262005 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.262046 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-576b5fdbd4-qmczw" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.263080 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.266231 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapic90a-account-delete-gg2x2" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.266950 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glancee859-account-delete-bb4lj" secret="" err="secret \"galera-openstack-dockercfg-n9fv6\" not found" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.298816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4e949a-2e3c-49ff-ba65-f925a9a95336-kube-api-access-hswgp" (OuterVolumeSpecName: "kube-api-access-hswgp") pod "3b4e949a-2e3c-49ff-ba65-f925a9a95336" (UID: "3b4e949a-2e3c-49ff-ba65-f925a9a95336"). InnerVolumeSpecName "kube-api-access-hswgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.300854 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.326207 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.355746 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data\") pod \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.356037 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data-custom\") pod \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.356092 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-scripts\") pod \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.358192 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-combined-ca-bundle\") pod \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.358254 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2tth\" (UniqueName: \"kubernetes.io/projected/deb8eeb7-c8f9-4324-a518-b837c3ad043c-kube-api-access-w2tth\") pod \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.358413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/deb8eeb7-c8f9-4324-a518-b837c3ad043c-etc-machine-id\") pod \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\" (UID: \"deb8eeb7-c8f9-4324-a518-b837c3ad043c\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.359966 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4e949a-2e3c-49ff-ba65-f925a9a95336-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.359993 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hswgp\" (UniqueName: \"kubernetes.io/projected/3b4e949a-2e3c-49ff-ba65-f925a9a95336-kube-api-access-hswgp\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.360608 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-scripts" (OuterVolumeSpecName: "scripts") pod "deb8eeb7-c8f9-4324-a518-b837c3ad043c" (UID: "deb8eeb7-c8f9-4324-a518-b837c3ad043c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.360922 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deb8eeb7-c8f9-4324-a518-b837c3ad043c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "deb8eeb7-c8f9-4324-a518-b837c3ad043c" (UID: "deb8eeb7-c8f9-4324-a518-b837c3ad043c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.362944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "deb8eeb7-c8f9-4324-a518-b837c3ad043c" (UID: "deb8eeb7-c8f9-4324-a518-b837c3ad043c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.363969 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb8eeb7-c8f9-4324-a518-b837c3ad043c-kube-api-access-w2tth" (OuterVolumeSpecName: "kube-api-access-w2tth") pod "deb8eeb7-c8f9-4324-a518-b837c3ad043c" (UID: "deb8eeb7-c8f9-4324-a518-b837c3ad043c"). InnerVolumeSpecName "kube-api-access-w2tth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.401895 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.435494 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.454823 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deb8eeb7-c8f9-4324-a518-b837c3ad043c" (UID: "deb8eeb7-c8f9-4324-a518-b837c3ad043c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.461281 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.465499 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.465531 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2tth\" (UniqueName: \"kubernetes.io/projected/deb8eeb7-c8f9-4324-a518-b837c3ad043c-kube-api-access-w2tth\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.465544 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/deb8eeb7-c8f9-4324-a518-b837c3ad043c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.465555 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.465567 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.480190 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.497911 4758 scope.go:117] "RemoveContainer" containerID="ef163a1a1695789c803e3b2391ac85959aee35c4dc735cf405bb6587fe19ed5f" Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.503547 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.504545 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.505327 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.505393 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.508211 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.517548 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.521874 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.539340 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.539941 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.539996 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.544940 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.569713 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-576b5fdbd4-qmczw"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.579695 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-576b5fdbd4-qmczw"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.585453 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data" (OuterVolumeSpecName: "config-data") pod "deb8eeb7-c8f9-4324-a518-b837c3ad043c" (UID: "deb8eeb7-c8f9-4324-a518-b837c3ad043c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.589414 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.612277 4758 scope.go:117] "RemoveContainer" containerID="b515931c8d15bc7880738113558415685d689f0ef349e841890a0dc9c902625e" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.643741 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.675055 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-config-data\") pod \"0754ea69-5007-4e51-bd22-0efd13954199\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.675157 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w2dg\" (UniqueName: \"kubernetes.io/projected/0754ea69-5007-4e51-bd22-0efd13954199-kube-api-access-9w2dg\") pod \"0754ea69-5007-4e51-bd22-0efd13954199\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.675449 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-combined-ca-bundle\") pod \"0754ea69-5007-4e51-bd22-0efd13954199\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.675474 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-memcached-tls-certs\") pod \"0754ea69-5007-4e51-bd22-0efd13954199\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.675499 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-kolla-config\") pod \"0754ea69-5007-4e51-bd22-0efd13954199\" (UID: \"0754ea69-5007-4e51-bd22-0efd13954199\") " Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.675992 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8eeb7-c8f9-4324-a518-b837c3ad043c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.676555 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "0754ea69-5007-4e51-bd22-0efd13954199" (UID: "0754ea69-5007-4e51-bd22-0efd13954199"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.677224 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-config-data" (OuterVolumeSpecName: "config-data") pod "0754ea69-5007-4e51-bd22-0efd13954199" (UID: "0754ea69-5007-4e51-bd22-0efd13954199"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.684393 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0754ea69-5007-4e51-bd22-0efd13954199-kube-api-access-9w2dg" (OuterVolumeSpecName: "kube-api-access-9w2dg") pod "0754ea69-5007-4e51-bd22-0efd13954199" (UID: "0754ea69-5007-4e51-bd22-0efd13954199"). InnerVolumeSpecName "kube-api-access-9w2dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.684411 4758 scope.go:117] "RemoveContainer" containerID="b0b70bf1d6c569af3a4f6fb6fd81f0660bf2b361866c0f71e0d8c87b60358a75" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.703887 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0754ea69-5007-4e51-bd22-0efd13954199" (UID: "0754ea69-5007-4e51-bd22-0efd13954199"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.732657 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "0754ea69-5007-4e51-bd22-0efd13954199" (UID: "0754ea69-5007-4e51-bd22-0efd13954199"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.739120 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" probeResult="failure" output="command timed out" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.779090 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w2dg\" (UniqueName: \"kubernetes.io/projected/0754ea69-5007-4e51-bd22-0efd13954199-kube-api-access-9w2dg\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.779137 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.779149 4758 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0754ea69-5007-4e51-bd22-0efd13954199-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.779161 4758 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.779174 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0754ea69-5007-4e51-bd22-0efd13954199-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.805669 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:23:12 crc kubenswrapper[4758]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Dec 03 17:23:12 crc kubenswrapper[4758]: > Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.815309 4758 scope.go:117] "RemoveContainer" containerID="0d525003c8cbf51daf8101b141ea9b8fbc2b99a11a1f0be921bbda88de524e41" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.860184 4758 scope.go:117] "RemoveContainer" containerID="6b35ae8e6d0ff4c3cd07edce5d23b46583d8b4af2ecf93f7e9477313990aa41c" Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.882387 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.882493 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:14.882470948 +0000 UTC m=+1650.083847799 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.882973 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.883014 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:23:14.883000891 +0000 UTC m=+1650.084377752 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.883088 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.883126 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:23:14.883117084 +0000 UTC m=+1650.084493945 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.883158 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.883177 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:23:14.883171276 +0000 UTC m=+1650.084548137 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.902969 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.926565 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.952093 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.970416 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-njckq"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.975283 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.977668 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 17:23:12 crc kubenswrapper[4758]: E1203 17:23:12.978247 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" Dec 03 17:23:12 crc kubenswrapper[4758]: I1203 17:23:12.987174 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-njckq"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.000211 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder14bb-account-delete-ctzv7"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.010471 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-14bb-account-create-update-s96dt"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.020891 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-14bb-account-create-update-s96dt"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.031209 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder14bb-account-delete-ctzv7"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.036550 4758 scope.go:117] "RemoveContainer" containerID="033147cfa4974a6e9865e1dce4a3ab80379ee25004c515e9a54cca5bc2a76e0e" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.068433 4758 scope.go:117] "RemoveContainer" containerID="a8202adddc1c7cf7ab92d5ed5e61b0495c0d800596e0b0472d95b872e5eca55f" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.119056 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-z5cp4"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.158078 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" path="/var/lib/kubelet/pods/04b3f0d2-7b01-48a6-b2be-9e205027d1a2/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.159078 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154d62cd-e6fa-4b15-ab82-7f486ea22eaa" path="/var/lib/kubelet/pods/154d62cd-e6fa-4b15-ab82-7f486ea22eaa/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.159592 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" path="/var/lib/kubelet/pods/1b49b1d3-cbb4-49da-8244-69e15595da33/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.160613 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" path="/var/lib/kubelet/pods/1f328daf-ac93-432b-8645-14b7030a24d1/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.162312 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b4e949a-2e3c-49ff-ba65-f925a9a95336" path="/var/lib/kubelet/pods/3b4e949a-2e3c-49ff-ba65-f925a9a95336/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.163263 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="428ee377-ea6d-4897-890a-6676a318c95a" path="/var/lib/kubelet/pods/428ee377-ea6d-4897-890a-6676a318c95a/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.164653 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" path="/var/lib/kubelet/pods/583229ad-ae2e-41b5-aad3-3c196b829f40/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.166620 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df131e1-f13f-4943-95a0-18086525f883" path="/var/lib/kubelet/pods/6df131e1-f13f-4943-95a0-18086525f883/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.167260 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" path="/var/lib/kubelet/pods/80b1161a-3f70-4956-b13a-1ef693aa2e7c/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.167767 4758 scope.go:117] "RemoveContainer" containerID="df8bc0faf1cabc23c62d9aea7423ca778a9372ef7f9e311140e90eae1b53b014" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.167910 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a1c0c2-3a27-43cc-8d41-d3c7508ab441" path="/var/lib/kubelet/pods/82a1c0c2-3a27-43cc-8d41-d3c7508ab441/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.169169 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84469de1-c706-4442-8610-31188889dad1" path="/var/lib/kubelet/pods/84469de1-c706-4442-8610-31188889dad1/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.170077 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" path="/var/lib/kubelet/pods/b461991e-58f6-4ff3-81d3-9cf2288d9b65/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: E1203 17:23:13.170322 4758 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 03 17:23:13 crc kubenswrapper[4758]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-03T17:23:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 03 17:23:13 crc kubenswrapper[4758]: /etc/init.d/functions: line 589: 484 Alarm clock "$@" Dec 03 17:23:13 crc kubenswrapper[4758]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-5czb6" message=< Dec 03 17:23:13 crc kubenswrapper[4758]: Exiting ovn-controller (1) [FAILED] Dec 03 17:23:13 crc kubenswrapper[4758]: Killing ovn-controller (1) [ OK ] Dec 03 17:23:13 crc kubenswrapper[4758]: Killing ovn-controller (1) with SIGKILL [ OK ] Dec 03 17:23:13 crc kubenswrapper[4758]: 2025-12-03T17:23:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 03 17:23:13 crc kubenswrapper[4758]: /etc/init.d/functions: line 589: 484 Alarm clock "$@" Dec 03 17:23:13 crc kubenswrapper[4758]: > Dec 03 17:23:13 crc kubenswrapper[4758]: E1203 17:23:13.170414 4758 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 03 17:23:13 crc kubenswrapper[4758]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-03T17:23:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 03 17:23:13 crc kubenswrapper[4758]: /etc/init.d/functions: line 589: 484 Alarm clock "$@" Dec 03 17:23:13 crc kubenswrapper[4758]: > pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" containerID="cri-o://12bf46a944f70cdab9ad1c78427dc024a221e2ca969030997c39eb4adbcaa629" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.170472 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-5czb6" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" containerID="cri-o://12bf46a944f70cdab9ad1c78427dc024a221e2ca969030997c39eb4adbcaa629" gracePeriod=22 Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.174734 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" path="/var/lib/kubelet/pods/be629420-0e78-44d4-9ca1-4ee1601d4ca6/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.176280 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" path="/var/lib/kubelet/pods/d0cc9a7c-fe92-4cfe-b8f6-a523db38b665/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.177301 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" path="/var/lib/kubelet/pods/deb8eeb7-c8f9-4324-a518-b837c3ad043c/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.178859 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7" path="/var/lib/kubelet/pods/ea7d4cc9-0b16-429d-a3b0-9a471c98a8f7/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.179441 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebee961b-48e5-4983-bb45-c14bbcc164f8" path="/var/lib/kubelet/pods/ebee961b-48e5-4983-bb45-c14bbcc164f8/volumes" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.180386 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-z5cp4"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.180451 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-02ea-account-create-update-drbs5"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.180467 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement02ea-account-delete-zbwhh"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.201753 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement02ea-account-delete-zbwhh"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.217771 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-02ea-account-create-update-drbs5"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.226030 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2kcmv"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.244397 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2kcmv"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.251434 4758 scope.go:117] "RemoveContainer" containerID="f6091694aad9cdd8c04fa3979ee4bf04648cfc927b2b56cd01f71c7f65050343" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.263734 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron988a-account-delete-t55hx"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.276956 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-988a-account-create-update-6fjck"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.291181 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron988a-account-delete-t55hx"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.301227 4758 scope.go:117] "RemoveContainer" containerID="9d2c93db5d7424078ddb1bb905ea651dea2607f1f94df3e14e8eaf9f6eec920f" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.303160 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-988a-account-create-update-6fjck"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.321935 4758 generic.go:334] "Generic (PLEG): container finished" podID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerID="aa23c937b63cdc7de84a3a65b1ab3cb567e5f8d745fd98201fbefbe82494cbc1" exitCode=0 Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.322062 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"070406cc-6d65-4fd0-81d7-264adead6bbd","Type":"ContainerDied","Data":"aa23c937b63cdc7de84a3a65b1ab3cb567e5f8d745fd98201fbefbe82494cbc1"} Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.331052 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5czb6_aac23ac3-72ea-49a7-9ded-6fabe28ff690/ovn-controller/0.log" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.331109 4758 generic.go:334] "Generic (PLEG): container finished" podID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerID="12bf46a944f70cdab9ad1c78427dc024a221e2ca969030997c39eb4adbcaa629" exitCode=137 Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.331246 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6" event={"ID":"aac23ac3-72ea-49a7-9ded-6fabe28ff690","Type":"ContainerDied","Data":"12bf46a944f70cdab9ad1c78427dc024a221e2ca969030997c39eb4adbcaa629"} Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.344951 4758 generic.go:334] "Generic (PLEG): container finished" podID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerID="ffded9f7e8bb729a9659af5430bde139dac85bb5e31acabdfd96c07c75f4d43d" exitCode=0 Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.345173 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e518390e-b4c4-49bb-913f-ddfc331456ab","Type":"ContainerDied","Data":"ffded9f7e8bb729a9659af5430bde139dac85bb5e31acabdfd96c07c75f4d43d"} Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.420904 4758 scope.go:117] "RemoveContainer" containerID="b86b8be6d1e76610577e86e34e754975aa88fa3ee8f7d13944550ef42f337ec8" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.440392 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0754ea69-5007-4e51-bd22-0efd13954199","Type":"ContainerDied","Data":"fc7a474a0f49cd7b9fbb633251ad467314a7636da93a401cf9ae22068a10a4e3"} Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.441129 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.457182 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qrldg"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.478960 4758 scope.go:117] "RemoveContainer" containerID="2071a07b9d8bd2fffbcfc63e0cd6e8e249aa95143b551ab078f950658b27507b" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.489248 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qrldg"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.570962 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicane6b7-account-delete-l2xbs"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.571280 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbicane6b7-account-delete-l2xbs" podUID="cb717c57-e263-449f-896f-38707d61e72c" containerName="mariadb-account-delete" containerID="cri-o://3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3" gracePeriod=30 Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.617150 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.680995 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e6b7-account-create-update-cxdn6"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684125 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-plugins-conf\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684175 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-confd\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684215 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684261 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/070406cc-6d65-4fd0-81d7-264adead6bbd-pod-info\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684337 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/070406cc-6d65-4fd0-81d7-264adead6bbd-erlang-cookie-secret\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684445 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-plugins\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684497 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-erlang-cookie\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684523 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-tls\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684546 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gk9r\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-kube-api-access-6gk9r\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-server-conf\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.684666 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data\") pod \"070406cc-6d65-4fd0-81d7-264adead6bbd\" (UID: \"070406cc-6d65-4fd0-81d7-264adead6bbd\") " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.692325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.710355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.713433 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.714230 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.721905 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.724749 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070406cc-6d65-4fd0-81d7-264adead6bbd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.737935 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/070406cc-6d65-4fd0-81d7-264adead6bbd-pod-info" (OuterVolumeSpecName: "pod-info") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.762773 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e6b7-account-create-update-cxdn6"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.778980 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-kube-api-access-6gk9r" (OuterVolumeSpecName: "kube-api-access-6gk9r") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "kube-api-access-6gk9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.784385 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data" (OuterVolumeSpecName: "config-data") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809177 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809226 4758 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/070406cc-6d65-4fd0-81d7-264adead6bbd-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809239 4758 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/070406cc-6d65-4fd0-81d7-264adead6bbd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809252 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809266 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809281 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809292 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gk9r\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-kube-api-access-6gk9r\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809306 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.809318 4758 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.875067 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.165:8776/healthcheck\": dial tcp 10.217.0.165:8776: i/o timeout" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.896125 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.901122 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.913262 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.928824 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.943024 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-frhtr"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.959349 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-frhtr"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.973284 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/keystone-584ffbf4c6-g7bkv" podUID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" containerName="keystone-api" probeResult="failure" output="Get \"https://10.217.0.152:5000/v3\": read tcp 10.217.0.2:47180->10.217.0.152:5000: read: connection reset by peer" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.989040 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-700b-account-create-update-8dvwz"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.989413 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-server-conf" (OuterVolumeSpecName: "server-conf") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.995426 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0700b-account-delete-snfrl"] Dec 03 17:23:13 crc kubenswrapper[4758]: I1203 17:23:13.995777 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell0700b-account-delete-snfrl" podUID="8a9b6914-42e5-4a9c-a4da-de3cee2587e2" containerName="mariadb-account-delete" containerID="cri-o://0cb9a87be8ad59f9ff8dd09d281da1a8164d0ffdaca0aabbf9f6081178c0baf1" gracePeriod=30 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.001906 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "070406cc-6d65-4fd0-81d7-264adead6bbd" (UID: "070406cc-6d65-4fd0-81d7-264adead6bbd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.004639 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-700b-account-create-update-8dvwz"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.015315 4758 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/070406cc-6d65-4fd0-81d7-264adead6bbd-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.015364 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/070406cc-6d65-4fd0-81d7-264adead6bbd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.060969 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-d4zrc"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.093491 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-d4zrc"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.135708 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c90a-account-create-update-24496"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.151834 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c90a-account-create-update-24496"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.159699 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic90a-account-delete-gg2x2"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.160118 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapic90a-account-delete-gg2x2" podUID="fef3d86b-78d4-4b18-9a05-1c2f435f03ed" containerName="mariadb-account-delete" containerID="cri-o://6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899" gracePeriod=30 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.168790 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-c92ds"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.178820 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-c92ds"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.183657 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e859-account-create-update-8k259"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.200342 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e859-account-create-update-8k259"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.202712 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancee859-account-delete-bb4lj"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.202998 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glancee859-account-delete-bb4lj" podUID="5a11fd25-44dc-4ed6-85a5-7f10fc42302a" containerName="mariadb-account-delete" containerID="cri-o://d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c" gracePeriod=30 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.236813 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.279695 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5czb6_aac23ac3-72ea-49a7-9ded-6fabe28ff690/ovn-controller/0.log" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.279798 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330618 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-combined-ca-bundle\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330696 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330737 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-confd\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330774 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8mjg\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-kube-api-access-g8mjg\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330803 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-ovn-controller-tls-certs\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330868 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e518390e-b4c4-49bb-913f-ddfc331456ab-pod-info\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-plugins\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.330992 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331065 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331109 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-erlang-cookie\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331141 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg224\" (UniqueName: \"kubernetes.io/projected/aac23ac3-72ea-49a7-9ded-6fabe28ff690-kube-api-access-rg224\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331220 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-tls\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331281 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-server-conf\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331359 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e518390e-b4c4-49bb-913f-ddfc331456ab-erlang-cookie-secret\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331386 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run-ovn\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331459 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331495 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-plugins-conf\") pod \"e518390e-b4c4-49bb-913f-ddfc331456ab\" (UID: \"e518390e-b4c4-49bb-913f-ddfc331456ab\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.331530 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-log-ovn\") pod \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\" (UID: \"aac23ac3-72ea-49a7-9ded-6fabe28ff690\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.332129 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.332181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run" (OuterVolumeSpecName: "var-run") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.333615 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.333819 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.334510 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.335427 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.336455 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts" (OuterVolumeSpecName: "scripts") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.339044 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.339404 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e518390e-b4c4-49bb-913f-ddfc331456ab-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.339528 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.352909 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.373855 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e518390e-b4c4-49bb-913f-ddfc331456ab-pod-info" (OuterVolumeSpecName: "pod-info") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.378051 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-kube-api-access-g8mjg" (OuterVolumeSpecName: "kube-api-access-g8mjg") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "kube-api-access-g8mjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.390802 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac23ac3-72ea-49a7-9ded-6fabe28ff690-kube-api-access-rg224" (OuterVolumeSpecName: "kube-api-access-rg224") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "kube-api-access-rg224". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.410382 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-server-conf" (OuterVolumeSpecName: "server-conf") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.430070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data" (OuterVolumeSpecName: "config-data") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433000 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-run-httpd\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433077 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcgwg\" (UniqueName: \"kubernetes.io/projected/7bb44365-2dda-4769-813a-712f362fff6b-kube-api-access-lcgwg\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433128 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-config-data\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433177 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-combined-ca-bundle\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433241 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-scripts\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433340 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-ceilometer-tls-certs\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433376 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-sg-core-conf-yaml\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433450 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-log-httpd\") pod \"7bb44365-2dda-4769-813a-712f362fff6b\" (UID: \"7bb44365-2dda-4769-813a-712f362fff6b\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.433941 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434178 4758 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e518390e-b4c4-49bb-913f-ddfc331456ab-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434194 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434208 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac23ac3-72ea-49a7-9ded-6fabe28ff690-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434253 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434266 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434277 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg224\" (UniqueName: \"kubernetes.io/projected/aac23ac3-72ea-49a7-9ded-6fabe28ff690-kube-api-access-rg224\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434286 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434295 4758 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434323 4758 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e518390e-b4c4-49bb-913f-ddfc331456ab-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434333 4758 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434342 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434351 4758 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e518390e-b4c4-49bb-913f-ddfc331456ab-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434359 4758 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434367 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434376 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac23ac3-72ea-49a7-9ded-6fabe28ff690-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434405 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8mjg\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-kube-api-access-g8mjg\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.434760 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.438096 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-scripts" (OuterVolumeSpecName: "scripts") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.442428 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb44365-2dda-4769-813a-712f362fff6b-kube-api-access-lcgwg" (OuterVolumeSpecName: "kube-api-access-lcgwg") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "kube-api-access-lcgwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.460030 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.484160 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.493231 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.497463 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aac23ac3-72ea-49a7-9ded-6fabe28ff690" (UID: "aac23ac3-72ea-49a7-9ded-6fabe28ff690"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.516382 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.535798 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data-custom\") pod \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.535919 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-combined-ca-bundle\") pod \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536038 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlsxt\" (UniqueName: \"kubernetes.io/projected/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-kube-api-access-dlsxt\") pod \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536132 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data\") pod \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536252 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-logs\") pod \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\" (UID: \"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536777 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536798 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536812 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bb44365-2dda-4769-813a-712f362fff6b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536824 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcgwg\" (UniqueName: \"kubernetes.io/projected/7bb44365-2dda-4769-813a-712f362fff6b-kube-api-access-lcgwg\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536837 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536849 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aac23ac3-72ea-49a7-9ded-6fabe28ff690-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.536859 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.537273 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-logs" (OuterVolumeSpecName: "logs") pod "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" (UID: "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.543945 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" (UID: "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.548374 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57b5ff9967-k6s52" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.549374 4758 generic.go:334] "Generic (PLEG): container finished" podID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerID="f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93" exitCode=0 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.549468 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57b5ff9967-k6s52" event={"ID":"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c","Type":"ContainerDied","Data":"f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.549505 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57b5ff9967-k6s52" event={"ID":"218264a1-c6a9-4ecb-8f40-5df1fd7dc77c","Type":"ContainerDied","Data":"48c44221348cf58dc317ed0742cadeb940a7fb2be358ee47a42d1c2dd8be5e84"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.549526 4758 scope.go:117] "RemoveContainer" containerID="f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.554594 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"070406cc-6d65-4fd0-81d7-264adead6bbd","Type":"ContainerDied","Data":"4d97514319b93120ec43aea8688d1f35b4a402fc2d7314bc55922c9f682aa8e1"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.554756 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.558063 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.560024 4758 generic.go:334] "Generic (PLEG): container finished" podID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" exitCode=0 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.560085 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b","Type":"ContainerDied","Data":"c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.565299 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-kube-api-access-dlsxt" (OuterVolumeSpecName: "kube-api-access-dlsxt") pod "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" (UID: "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c"). InnerVolumeSpecName "kube-api-access-dlsxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.565354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e518390e-b4c4-49bb-913f-ddfc331456ab","Type":"ContainerDied","Data":"184e063513f2f136edb7bc9d67f4581f9420630c7e2d1fd923601f13ea501aa5"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.565583 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.574037 4758 generic.go:334] "Generic (PLEG): container finished" podID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" containerID="1d73cb8d75674ff7493a3d6896b8923851258760396fa1c116f8e006497ff57c" exitCode=0 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.574125 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-584ffbf4c6-g7bkv" event={"ID":"6bdefd9e-77f0-4137-8b30-35aca9f17c3f","Type":"ContainerDied","Data":"1d73cb8d75674ff7493a3d6896b8923851258760396fa1c116f8e006497ff57c"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.577834 4758 generic.go:334] "Generic (PLEG): container finished" podID="7bb44365-2dda-4769-813a-712f362fff6b" containerID="d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c" exitCode=0 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.577897 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerDied","Data":"d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.577924 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bb44365-2dda-4769-813a-712f362fff6b","Type":"ContainerDied","Data":"224ac65b99db035304592b13991230b1bbf407cc7a5745c110138c8c18b92e9e"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.578017 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.584115 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.586137 4758 generic.go:334] "Generic (PLEG): container finished" podID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerID="efb795a450acef7b396db339028440c950cbfa2f32b9d7f9c051e83e71c85fb6" exitCode=0 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.586222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" event={"ID":"53b74dc9-5836-4c5c-bae8-803f64c9de12","Type":"ContainerDied","Data":"efb795a450acef7b396db339028440c950cbfa2f32b9d7f9c051e83e71c85fb6"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.590201 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5czb6_aac23ac3-72ea-49a7-9ded-6fabe28ff690/ovn-controller/0.log" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.590341 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5czb6" event={"ID":"aac23ac3-72ea-49a7-9ded-6fabe28ff690","Type":"ContainerDied","Data":"5254ee401bfb1f603de824aaac045083787e69c235fdd2b1fd80e2eb768d747f"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.590510 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5czb6" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.592154 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" (UID: "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.595034 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_daf64139-0696-484a-870f-023898fbeb08/ovn-northd/0.log" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.595082 4758 generic.go:334] "Generic (PLEG): container finished" podID="daf64139-0696-484a-870f-023898fbeb08" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" exitCode=139 Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.595147 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"daf64139-0696-484a-870f-023898fbeb08","Type":"ContainerDied","Data":"c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7"} Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.599568 4758 scope.go:117] "RemoveContainer" containerID="7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.618094 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e518390e-b4c4-49bb-913f-ddfc331456ab" (UID: "e518390e-b4c4-49bb-913f-ddfc331456ab"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.623538 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.628266 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637848 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637865 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637874 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637884 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlsxt\" (UniqueName: \"kubernetes.io/projected/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-kube-api-access-dlsxt\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637892 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637901 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e518390e-b4c4-49bb-913f-ddfc331456ab-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.637909 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.646007 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-config-data" (OuterVolumeSpecName: "config-data") pod "7bb44365-2dda-4769-813a-712f362fff6b" (UID: "7bb44365-2dda-4769-813a-712f362fff6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.684599 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data" (OuterVolumeSpecName: "config-data") pod "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" (UID: "218264a1-c6a9-4ecb-8f40-5df1fd7dc77c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.740028 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.740072 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb44365-2dda-4769-813a-712f362fff6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.855735 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.868573 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5czb6"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.869429 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_daf64139-0696-484a-870f-023898fbeb08/ovn-northd/0.log" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.869526 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.875107 4758 scope.go:117] "RemoveContainer" containerID="f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93" Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.888986 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93\": container with ID starting with f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93 not found: ID does not exist" containerID="f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.889120 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93"} err="failed to get container status \"f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93\": rpc error: code = NotFound desc = could not find container \"f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93\": container with ID starting with f76da1018dde53af627799be93620db78bb2eb86f8e1b40501af027690679e93 not found: ID does not exist" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.889195 4758 scope.go:117] "RemoveContainer" containerID="7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4" Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.901186 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4\": container with ID starting with 7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4 not found: ID does not exist" containerID="7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.901249 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4"} err="failed to get container status \"7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4\": rpc error: code = NotFound desc = could not find container \"7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4\": container with ID starting with 7272a007cb99a0fc944d20e04d0ee4c5642e5cb382ae5c59174121a41a059bc4 not found: ID does not exist" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.901291 4758 scope.go:117] "RemoveContainer" containerID="aa23c937b63cdc7de84a3a65b1ab3cb567e5f8d745fd98201fbefbe82494cbc1" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.938132 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5czb6"] Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.944198 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-combined-ca-bundle\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.944527 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-ovn-northd-tls-certs\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.944766 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbmxq\" (UniqueName: \"kubernetes.io/projected/daf64139-0696-484a-870f-023898fbeb08-kube-api-access-nbmxq\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.944995 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b74dc9-5836-4c5c-bae8-803f64c9de12-logs\") pod \"53b74dc9-5836-4c5c-bae8-803f64c9de12\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.945227 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-scripts\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.945445 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data\") pod \"53b74dc9-5836-4c5c-bae8-803f64c9de12\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.945619 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfhqs\" (UniqueName: \"kubernetes.io/projected/53b74dc9-5836-4c5c-bae8-803f64c9de12-kube-api-access-sfhqs\") pod \"53b74dc9-5836-4c5c-bae8-803f64c9de12\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.945784 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-config\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.945883 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daf64139-0696-484a-870f-023898fbeb08-ovn-rundir\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.946618 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data-custom\") pod \"53b74dc9-5836-4c5c-bae8-803f64c9de12\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.947100 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-metrics-certs-tls-certs\") pod \"daf64139-0696-484a-870f-023898fbeb08\" (UID: \"daf64139-0696-484a-870f-023898fbeb08\") " Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.947307 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-combined-ca-bundle\") pod \"53b74dc9-5836-4c5c-bae8-803f64c9de12\" (UID: \"53b74dc9-5836-4c5c-bae8-803f64c9de12\") " Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.949864 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.950066 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:23:18.950038398 +0000 UTC m=+1654.151415259 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.954057 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.954177 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:23:18.954144328 +0000 UTC m=+1654.155521189 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.954252 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.954293 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:23:18.954277421 +0000 UTC m=+1654.155654282 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.955005 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.955029 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daf64139-0696-484a-870f-023898fbeb08-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: E1203 17:23:14.955116 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:18.955085113 +0000 UTC m=+1654.156461974 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.955713 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b74dc9-5836-4c5c-bae8-803f64c9de12-kube-api-access-sfhqs" (OuterVolumeSpecName: "kube-api-access-sfhqs") pod "53b74dc9-5836-4c5c-bae8-803f64c9de12" (UID: "53b74dc9-5836-4c5c-bae8-803f64c9de12"). InnerVolumeSpecName "kube-api-access-sfhqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.957604 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-config" (OuterVolumeSpecName: "config") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.966911 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf64139-0696-484a-870f-023898fbeb08-kube-api-access-nbmxq" (OuterVolumeSpecName: "kube-api-access-nbmxq") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "kube-api-access-nbmxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.969672 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "53b74dc9-5836-4c5c-bae8-803f64c9de12" (UID: "53b74dc9-5836-4c5c-bae8-803f64c9de12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.972615 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-scripts" (OuterVolumeSpecName: "scripts") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:14 crc kubenswrapper[4758]: I1203 17:23:14.979206 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.000457 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53b74dc9-5836-4c5c-bae8-803f64c9de12" (UID: "53b74dc9-5836-4c5c-bae8-803f64c9de12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.004148 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.006483 4758 scope.go:117] "RemoveContainer" containerID="6de2813edf252cbe7e4d319d670dd48259def0afd5b8b085c56a64c56604dd50" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.009751 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.032729 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.043468 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.045351 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-57b5ff9967-k6s52"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.048059 4758 scope.go:117] "RemoveContainer" containerID="ffded9f7e8bb729a9659af5430bde139dac85bb5e31acabdfd96c07c75f4d43d" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.055271 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53b74dc9-5836-4c5c-bae8-803f64c9de12-logs" (OuterVolumeSpecName: "logs") pod "53b74dc9-5836-4c5c-bae8-803f64c9de12" (UID: "53b74dc9-5836-4c5c-bae8-803f64c9de12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.060082 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-57b5ff9967-k6s52"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062238 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-combined-ca-bundle\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062319 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-public-tls-certs\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-config-data\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062374 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-fernet-keys\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062430 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-internal-tls-certs\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062447 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-scripts\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062493 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-924wf\" (UniqueName: \"kubernetes.io/projected/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-kube-api-access-924wf\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.062561 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-credential-keys\") pod \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\" (UID: \"6bdefd9e-77f0-4137-8b30-35aca9f17c3f\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.063829 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.065629 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data" (OuterVolumeSpecName: "config-data") pod "53b74dc9-5836-4c5c-bae8-803f64c9de12" (UID: "53b74dc9-5836-4c5c-bae8-803f64c9de12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.074562 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-kube-api-access-924wf" (OuterVolumeSpecName: "kube-api-access-924wf") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "kube-api-access-924wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079476 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079520 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfhqs\" (UniqueName: \"kubernetes.io/projected/53b74dc9-5836-4c5c-bae8-803f64c9de12-kube-api-access-sfhqs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079539 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079548 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/daf64139-0696-484a-870f-023898fbeb08-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079557 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079571 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b74dc9-5836-4c5c-bae8-803f64c9de12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079581 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079590 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079601 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbmxq\" (UniqueName: \"kubernetes.io/projected/daf64139-0696-484a-870f-023898fbeb08-kube-api-access-nbmxq\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079613 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-924wf\" (UniqueName: \"kubernetes.io/projected/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-kube-api-access-924wf\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079622 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b74dc9-5836-4c5c-bae8-803f64c9de12-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.079633 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf64139-0696-484a-870f-023898fbeb08-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.089440 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.092371 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-scripts" (OuterVolumeSpecName: "scripts") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.100119 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.164061 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.164278 4758 scope.go:117] "RemoveContainer" containerID="085ad6a4f1116fe1014b94a49396f4beca5e6b5a210c06edc52bb5e34006f632" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.183062 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-config-data\") pod \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.183115 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-combined-ca-bundle\") pod \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.183233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7rhz\" (UniqueName: \"kubernetes.io/projected/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-kube-api-access-r7rhz\") pod \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\" (UID: \"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.183667 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.183708 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.186745 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.193102 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.212914 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "daf64139-0696-484a-870f-023898fbeb08" (UID: "daf64139-0696-484a-870f-023898fbeb08"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.230178 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="014153c0-72a5-46ae-9065-ff34c290745a" path="/var/lib/kubelet/pods/014153c0-72a5-46ae-9065-ff34c290745a/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.230994 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" path="/var/lib/kubelet/pods/070406cc-6d65-4fd0-81d7-264adead6bbd/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.231601 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0754ea69-5007-4e51-bd22-0efd13954199" path="/var/lib/kubelet/pods/0754ea69-5007-4e51-bd22-0efd13954199/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.256369 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a0e71a2-24ae-4842-935e-5ed7a9bc7145" path="/var/lib/kubelet/pods/0a0e71a2-24ae-4842-935e-5ed7a9bc7145/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.257000 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20469a6a-4abc-4e0b-886f-ba199ee5caa3" path="/var/lib/kubelet/pods/20469a6a-4abc-4e0b-886f-ba199ee5caa3/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.257593 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" path="/var/lib/kubelet/pods/218264a1-c6a9-4ecb-8f40-5df1fd7dc77c/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.287704 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22fa959d-5438-45b3-a053-a86a205e8da4" path="/var/lib/kubelet/pods/22fa959d-5438-45b3-a053-a86a205e8da4/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.288257 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401d7406-8683-47aa-9562-5d3ee99ee30e" path="/var/lib/kubelet/pods/401d7406-8683-47aa-9562-5d3ee99ee30e/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.289930 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf64139-0696-484a-870f-023898fbeb08-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.289953 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.289962 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.292935 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-kube-api-access-r7rhz" (OuterVolumeSpecName: "kube-api-access-r7rhz") pod "eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" (UID: "eeb5db84-3c14-46e1-a18b-2a0ac35ea34b"). InnerVolumeSpecName "kube-api-access-r7rhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.305574 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb44365-2dda-4769-813a-712f362fff6b" path="/var/lib/kubelet/pods/7bb44365-2dda-4769-813a-712f362fff6b/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.306701 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff4c92a-13c5-4a18-944d-38682d75ec07" path="/var/lib/kubelet/pods/7ff4c92a-13c5-4a18-944d-38682d75ec07/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.307848 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97d849bd-1f04-4a59-805a-cb9ae0f3b7bd" path="/var/lib/kubelet/pods/97d849bd-1f04-4a59-805a-cb9ae0f3b7bd/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.308454 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a46406f0-070f-4bb5-890c-50d67da0a338" path="/var/lib/kubelet/pods/a46406f0-070f-4bb5-890c-50d67da0a338/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.309299 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" path="/var/lib/kubelet/pods/aac23ac3-72ea-49a7-9ded-6fabe28ff690/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.309925 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8eb3ef8-1b17-4d2d-b664-f32d7263d559" path="/var/lib/kubelet/pods/b8eb3ef8-1b17-4d2d-b664-f32d7263d559/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.319416 4758 scope.go:117] "RemoveContainer" containerID="922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.321930 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.322765 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde9ec06-c65f-45a5-9e90-5a12397f010e" path="/var/lib/kubelet/pods/bde9ec06-c65f-45a5-9e90-5a12397f010e/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.323287 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d868cfc0-6d71-4170-b50d-5f6a411111c8" path="/var/lib/kubelet/pods/d868cfc0-6d71-4170-b50d-5f6a411111c8/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.323825 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2e397d-cdcc-4116-8dfb-5b3ac6da924c" path="/var/lib/kubelet/pods/dc2e397d-cdcc-4116-8dfb-5b3ac6da924c/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.324938 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e27b94dc-9078-4c4c-911b-1d1501e71734" path="/var/lib/kubelet/pods/e27b94dc-9078-4c4c-911b-1d1501e71734/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.325778 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" path="/var/lib/kubelet/pods/e518390e-b4c4-49bb-913f-ddfc331456ab/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.326522 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fde0a172-de16-4feb-a2e0-c51ccb8f41e2" path="/var/lib/kubelet/pods/fde0a172-de16-4feb-a2e0-c51ccb8f41e2/volumes" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.348921 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-config-data" (OuterVolumeSpecName: "config-data") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.379488 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-config-data" (OuterVolumeSpecName: "config-data") pod "eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" (UID: "eeb5db84-3c14-46e1-a18b-2a0ac35ea34b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.392429 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.392471 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7rhz\" (UniqueName: \"kubernetes.io/projected/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-kube-api-access-r7rhz\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.392506 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.392516 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.396900 4758 scope.go:117] "RemoveContainer" containerID="a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.427668 4758 scope.go:117] "RemoveContainer" containerID="d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.432888 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" (UID: "eeb5db84-3c14-46e1-a18b-2a0ac35ea34b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.437219 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6bdefd9e-77f0-4137-8b30-35aca9f17c3f" (UID: "6bdefd9e-77f0-4137-8b30-35aca9f17c3f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.456432 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-576b5fdbd4-qmczw" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.457612 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-576b5fdbd4-qmczw" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.478782 4758 scope.go:117] "RemoveContainer" containerID="e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.494478 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bdefd9e-77f0-4137-8b30-35aca9f17c3f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.494582 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.550434 4758 scope.go:117] "RemoveContainer" containerID="922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a" Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.551383 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a\": container with ID starting with 922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a not found: ID does not exist" containerID="922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.551444 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a"} err="failed to get container status \"922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a\": rpc error: code = NotFound desc = could not find container \"922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a\": container with ID starting with 922001a26da4cdd729b11130843bad6aeb7152fdc3bc4d5ac09a3db252aec97a not found: ID does not exist" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.551488 4758 scope.go:117] "RemoveContainer" containerID="a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad" Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.553050 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad\": container with ID starting with a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad not found: ID does not exist" containerID="a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.553083 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad"} err="failed to get container status \"a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad\": rpc error: code = NotFound desc = could not find container \"a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad\": container with ID starting with a3a0fdf4f00077500badd513b81cf2b551b58e137918e2002d21833f4e3b0fad not found: ID does not exist" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.553104 4758 scope.go:117] "RemoveContainer" containerID="d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c" Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.553523 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c\": container with ID starting with d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c not found: ID does not exist" containerID="d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.553583 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c"} err="failed to get container status \"d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c\": rpc error: code = NotFound desc = could not find container \"d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c\": container with ID starting with d6ea0068e6d38200820d287110564e21f7362e2fd9318192f7e5ae8410b1303c not found: ID does not exist" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.553626 4758 scope.go:117] "RemoveContainer" containerID="e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388" Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.554013 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388\": container with ID starting with e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388 not found: ID does not exist" containerID="e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.554042 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388"} err="failed to get container status \"e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388\": rpc error: code = NotFound desc = could not find container \"e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388\": container with ID starting with e41b23f8e62a43921b3863b8c67d19ac047a20bb6eb441053e385304c19b8388 not found: ID does not exist" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.554056 4758 scope.go:117] "RemoveContainer" containerID="12bf46a944f70cdab9ad1c78427dc024a221e2ca969030997c39eb4adbcaa629" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.650987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" event={"ID":"53b74dc9-5836-4c5c-bae8-803f64c9de12","Type":"ContainerDied","Data":"89ab0ca141fa024b4e4a491501e9e620f3075821ae28fae8bcfdd12d52ef5bbd"} Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.651055 4758 scope.go:117] "RemoveContainer" containerID="efb795a450acef7b396db339028440c950cbfa2f32b9d7f9c051e83e71c85fb6" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.651254 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bb9966db6-phk6t" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.656914 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eeb5db84-3c14-46e1-a18b-2a0ac35ea34b","Type":"ContainerDied","Data":"94664d5ac08789bb46986d1262ce4f8a9b119515f5d6a7d6c618257f02119fdc"} Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.657060 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.665695 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-584ffbf4c6-g7bkv" event={"ID":"6bdefd9e-77f0-4137-8b30-35aca9f17c3f","Type":"ContainerDied","Data":"be4a4e6fb13c4e45b0dc305c4994a606ae4bcc366eabeac00971b054e1a8939c"} Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.665831 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-584ffbf4c6-g7bkv" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.689391 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_daf64139-0696-484a-870f-023898fbeb08/ovn-northd/0.log" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.689516 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"daf64139-0696-484a-870f-023898fbeb08","Type":"ContainerDied","Data":"f1c7d1ad6d07d74047a1d1601a9af41bcc9dba404849ff99f0561d54cb0315b9"} Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.689652 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.721360 4758 scope.go:117] "RemoveContainer" containerID="9c02a3775a91ddea9562f2c00063105c43fac3fa4443572b50d1f8eb0a7c2803" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.722534 4758 generic.go:334] "Generic (PLEG): container finished" podID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerID="d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea" exitCode=0 Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.722608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9c80de0e-2bee-4e3f-8edf-f4a235b01d76","Type":"ContainerDied","Data":"d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea"} Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.737820 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7bb9966db6-phk6t"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.748192 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7bb9966db6-phk6t"] Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.760879 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea is running failed: container process not found" containerID="d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.761493 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea is running failed: container process not found" containerID="d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.761711 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea is running failed: container process not found" containerID="d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 03 17:23:15 crc kubenswrapper[4758]: E1203 17:23:15.761747 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="galera" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.763513 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.769523 4758 scope.go:117] "RemoveContainer" containerID="c1d1bbde152d81bac732c4c88c6d3910d8567ccc2e8ee75b3281b94274a45726" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.775563 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.785504 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-584ffbf4c6-g7bkv"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.796160 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-584ffbf4c6-g7bkv"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.799525 4758 scope.go:117] "RemoveContainer" containerID="1d73cb8d75674ff7493a3d6896b8923851258760396fa1c116f8e006497ff57c" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.806852 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.815504 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.875781 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.892601 4758 scope.go:117] "RemoveContainer" containerID="43d7c9ec7f4e1c689ef29fc35a5e148631fe6789db20d2476a17525c48bb0c03" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.902459 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-operator-scripts\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.902547 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.902619 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-combined-ca-bundle\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.902825 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccv6m\" (UniqueName: \"kubernetes.io/projected/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kube-api-access-ccv6m\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.902942 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-galera-tls-certs\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.903003 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-generated\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.903549 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.903761 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-default\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.903790 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kolla-config\") pod \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\" (UID: \"9c80de0e-2bee-4e3f-8edf-f4a235b01d76\") " Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.904309 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.908282 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.908311 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.909056 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.909783 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kube-api-access-ccv6m" (OuterVolumeSpecName: "kube-api-access-ccv6m") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "kube-api-access-ccv6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.913458 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.940856 4758 scope.go:117] "RemoveContainer" containerID="c64e2813eabe7531b5a7b02ce0839377c6808670573cff4ccb7a803af1339eb7" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.946695 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:15 crc kubenswrapper[4758]: I1203 17:23:15.969248 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "9c80de0e-2bee-4e3f-8edf-f4a235b01d76" (UID: "9c80de0e-2bee-4e3f-8edf-f4a235b01d76"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006183 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccv6m\" (UniqueName: \"kubernetes.io/projected/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kube-api-access-ccv6m\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006227 4758 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006239 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006251 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006263 4758 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006295 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.006310 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80de0e-2bee-4e3f-8edf-f4a235b01d76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.025810 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.108398 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:16 crc kubenswrapper[4758]: E1203 17:23:16.341637 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/system-systemd\\\\x2dcoredump.slice/systemd-coredump@1-78198-0.service\": RecentStats: unable to find data in memory cache]" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.762275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9c80de0e-2bee-4e3f-8edf-f4a235b01d76","Type":"ContainerDied","Data":"9ba0dbb96f7bfc52b8dc940d806620d01eb3d778c6e4b13f0e71b9a472076e1e"} Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.762863 4758 scope.go:117] "RemoveContainer" containerID="d9b253106eb77186c982caf42fbd6b790d5ec7ef317ee7b9ff7ea55d627910ea" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.763085 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.814529 4758 scope.go:117] "RemoveContainer" containerID="ef2a638cba2c7b4383584dce02eb1679141386353eb7c10bd4bd9af144b23c05" Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.829375 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:23:16 crc kubenswrapper[4758]: I1203 17:23:16.837666 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:23:17 crc kubenswrapper[4758]: I1203 17:23:17.144147 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" path="/var/lib/kubelet/pods/53b74dc9-5836-4c5c-bae8-803f64c9de12/volumes" Dec 03 17:23:17 crc kubenswrapper[4758]: I1203 17:23:17.144820 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" path="/var/lib/kubelet/pods/6bdefd9e-77f0-4137-8b30-35aca9f17c3f/volumes" Dec 03 17:23:17 crc kubenswrapper[4758]: I1203 17:23:17.145443 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" path="/var/lib/kubelet/pods/9c80de0e-2bee-4e3f-8edf-f4a235b01d76/volumes" Dec 03 17:23:17 crc kubenswrapper[4758]: I1203 17:23:17.146763 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf64139-0696-484a-870f-023898fbeb08" path="/var/lib/kubelet/pods/daf64139-0696-484a-870f-023898fbeb08/volumes" Dec 03 17:23:17 crc kubenswrapper[4758]: I1203 17:23:17.147398 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" path="/var/lib/kubelet/pods/eeb5db84-3c14-46e1-a18b-2a0ac35ea34b/volumes" Dec 03 17:23:17 crc kubenswrapper[4758]: I1203 17:23:17.301923 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="0754ea69-5007-4e51-bd22-0efd13954199" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.106:11211: i/o timeout" Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.503795 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.504218 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.504466 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.504510 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.505922 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.507121 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.508350 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:17 crc kubenswrapper[4758]: E1203 17:23:17.508385 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.978712 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.979212 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:23:26.979188249 +0000 UTC m=+1662.180565110 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.978712 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.979659 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:23:26.979651631 +0000 UTC m=+1662.181028492 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.978757 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.979713 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:23:26.979704292 +0000 UTC m=+1662.181081153 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.978790 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:18 crc kubenswrapper[4758]: E1203 17:23:18.979742 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:26.979736293 +0000 UTC m=+1662.181113154 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:19 crc kubenswrapper[4758]: I1203 17:23:19.119099 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:23:19 crc kubenswrapper[4758]: E1203 17:23:19.120070 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.503353 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.504604 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.505082 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.505957 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.506003 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.507977 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.510504 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:22 crc kubenswrapper[4758]: E1203 17:23:22.510617 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:23:23 crc kubenswrapper[4758]: I1203 17:23:23.862828 4758 generic.go:334] "Generic (PLEG): container finished" podID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerID="2496f74c6aa8a3c995f6fcb782d022aa7cb8ab73e8224330eecc746a01760d7c" exitCode=0 Dec 03 17:23:23 crc kubenswrapper[4758]: I1203 17:23:23.862918 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dc74dcbf-wzmb5" event={"ID":"bd709657-fb8d-4c81-873f-a0b44d63abef","Type":"ContainerDied","Data":"2496f74c6aa8a3c995f6fcb782d022aa7cb8ab73e8224330eecc746a01760d7c"} Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.014033 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083493 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-combined-ca-bundle\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083602 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9vdz\" (UniqueName: \"kubernetes.io/projected/bd709657-fb8d-4c81-873f-a0b44d63abef-kube-api-access-c9vdz\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083633 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-public-tls-certs\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083720 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-httpd-config\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083794 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-ovndb-tls-certs\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083819 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-config\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.083903 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-internal-tls-certs\") pod \"bd709657-fb8d-4c81-873f-a0b44d63abef\" (UID: \"bd709657-fb8d-4c81-873f-a0b44d63abef\") " Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.090596 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd709657-fb8d-4c81-873f-a0b44d63abef-kube-api-access-c9vdz" (OuterVolumeSpecName: "kube-api-access-c9vdz") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "kube-api-access-c9vdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.091545 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.134543 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.136318 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-config" (OuterVolumeSpecName: "config") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.142547 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.145113 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.149496 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "bd709657-fb8d-4c81-873f-a0b44d63abef" (UID: "bd709657-fb8d-4c81-873f-a0b44d63abef"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187032 4758 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187071 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187084 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187095 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187104 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9vdz\" (UniqueName: \"kubernetes.io/projected/bd709657-fb8d-4c81-873f-a0b44d63abef-kube-api-access-c9vdz\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187115 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.187123 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd709657-fb8d-4c81-873f-a0b44d63abef-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.877494 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dc74dcbf-wzmb5" event={"ID":"bd709657-fb8d-4c81-873f-a0b44d63abef","Type":"ContainerDied","Data":"6083b0d716f3869aee6fc96831264050489b32214dcf21d56e722d859870aea3"} Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.878199 4758 scope.go:117] "RemoveContainer" containerID="6637a67a10c678e5c8c666412014de74c5391a41c691d978da31902186b3e7c0" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.877604 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dc74dcbf-wzmb5" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.922796 4758 scope.go:117] "RemoveContainer" containerID="2496f74c6aa8a3c995f6fcb782d022aa7cb8ab73e8224330eecc746a01760d7c" Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.929158 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69dc74dcbf-wzmb5"] Dec 03 17:23:24 crc kubenswrapper[4758]: I1203 17:23:24.937633 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69dc74dcbf-wzmb5"] Dec 03 17:23:25 crc kubenswrapper[4758]: I1203 17:23:25.126008 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" path="/var/lib/kubelet/pods/bd709657-fb8d-4c81-873f-a0b44d63abef/volumes" Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044130 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044660 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:23:43.044633911 +0000 UTC m=+1678.246010772 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044130 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044831 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:23:43.044801107 +0000 UTC m=+1678.246178078 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044218 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044959 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:23:43.04493371 +0000 UTC m=+1678.246310571 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.044243 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.045040 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:23:43.045024443 +0000 UTC m=+1678.246401484 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.503621 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.504393 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.505370 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.505462 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.505494 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.508132 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.510188 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:27 crc kubenswrapper[4758]: E1203 17:23:27.510251 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:23:30 crc kubenswrapper[4758]: I1203 17:23:30.114702 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:23:30 crc kubenswrapper[4758]: E1203 17:23:30.115733 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.504131 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.505071 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.505572 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.505614 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.506274 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.508359 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.510251 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 17:23:32 crc kubenswrapper[4758]: E1203 17:23:32.510290 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-5mttb" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.667344 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.710351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vvcx\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-kube-api-access-2vvcx\") pod \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.710442 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.710486 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-cache\") pod \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.710598 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") pod \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.710808 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-lock\") pod \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\" (UID: \"946a9b5f-7abf-4b3e-bc51-08a45e73ff30\") " Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.711753 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-lock" (OuterVolumeSpecName: "lock") pod "946a9b5f-7abf-4b3e-bc51-08a45e73ff30" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.712164 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-cache" (OuterVolumeSpecName: "cache") pod "946a9b5f-7abf-4b3e-bc51-08a45e73ff30" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.718209 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "946a9b5f-7abf-4b3e-bc51-08a45e73ff30" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.718997 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "946a9b5f-7abf-4b3e-bc51-08a45e73ff30" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.720379 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-kube-api-access-2vvcx" (OuterVolumeSpecName: "kube-api-access-2vvcx") pod "946a9b5f-7abf-4b3e-bc51-08a45e73ff30" (UID: "946a9b5f-7abf-4b3e-bc51-08a45e73ff30"). InnerVolumeSpecName "kube-api-access-2vvcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.812993 4758 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-lock\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.813044 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vvcx\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-kube-api-access-2vvcx\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.813085 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.813098 4758 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-cache\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.813111 4758 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/946a9b5f-7abf-4b3e-bc51-08a45e73ff30-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.828947 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.915481 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.952357 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5mttb_3e30b3f9-d35f-43a0-9f06-25ebf09ccb34/ovs-vswitchd/0.log" Dec 03 17:23:35 crc kubenswrapper[4758]: I1203 17:23:35.954259 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.015481 4758 generic.go:334] "Generic (PLEG): container finished" podID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerID="1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27" exitCode=137 Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.015585 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27"} Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.015625 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.015660 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"946a9b5f-7abf-4b3e-bc51-08a45e73ff30","Type":"ContainerDied","Data":"807f57fce17f9d09496cb0a6f17fc8ab10c2dd9c1146877fa4fa1db6de9554ba"} Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.015710 4758 scope.go:117] "RemoveContainer" containerID="1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.016872 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nb2n\" (UniqueName: \"kubernetes.io/projected/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-kube-api-access-2nb2n\") pod \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.016967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-run\") pod \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017024 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-scripts\") pod \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017063 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-run" (OuterVolumeSpecName: "var-run") pod "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" (UID: "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017067 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-log\") pod \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017119 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-etc-ovs\") pod \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017148 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-lib\") pod \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\" (UID: \"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34\") " Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017199 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-log" (OuterVolumeSpecName: "var-log") pod "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" (UID: "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017239 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" (UID: "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017563 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-lib" (OuterVolumeSpecName: "var-lib") pod "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" (UID: "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017645 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017664 4758 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.017675 4758 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.019332 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-scripts" (OuterVolumeSpecName: "scripts") pod "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" (UID: "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.019365 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5mttb_3e30b3f9-d35f-43a0-9f06-25ebf09ccb34/ovs-vswitchd/0.log" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.021101 4758 generic.go:334] "Generic (PLEG): container finished" podID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" exitCode=137 Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.021150 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerDied","Data":"e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70"} Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.021199 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5mttb" event={"ID":"3e30b3f9-d35f-43a0-9f06-25ebf09ccb34","Type":"ContainerDied","Data":"2692fbc5fd698183cfebe16cf94c0a605e5e604f7a69a78096bf4fd2a8c9c8d0"} Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.021290 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5mttb" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.021859 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-kube-api-access-2nb2n" (OuterVolumeSpecName: "kube-api-access-2nb2n") pod "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" (UID: "3e30b3f9-d35f-43a0-9f06-25ebf09ccb34"). InnerVolumeSpecName "kube-api-access-2nb2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.053593 4758 scope.go:117] "RemoveContainer" containerID="fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.063143 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.068663 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.076814 4758 scope.go:117] "RemoveContainer" containerID="1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.096822 4758 scope.go:117] "RemoveContainer" containerID="1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.119335 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nb2n\" (UniqueName: \"kubernetes.io/projected/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-kube-api-access-2nb2n\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.119380 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.119394 4758 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34-var-lib\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.123407 4758 scope.go:117] "RemoveContainer" containerID="187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.147809 4758 scope.go:117] "RemoveContainer" containerID="98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.173915 4758 scope.go:117] "RemoveContainer" containerID="5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.197367 4758 scope.go:117] "RemoveContainer" containerID="a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.254339 4758 scope.go:117] "RemoveContainer" containerID="4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.276139 4758 scope.go:117] "RemoveContainer" containerID="48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.294818 4758 scope.go:117] "RemoveContainer" containerID="dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.315751 4758 scope.go:117] "RemoveContainer" containerID="7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.352903 4758 scope.go:117] "RemoveContainer" containerID="8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.364634 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-5mttb"] Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.373936 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-5mttb"] Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.382282 4758 scope.go:117] "RemoveContainer" containerID="93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.414823 4758 scope.go:117] "RemoveContainer" containerID="def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.440500 4758 scope.go:117] "RemoveContainer" containerID="1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.441495 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27\": container with ID starting with 1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27 not found: ID does not exist" containerID="1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.441749 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27"} err="failed to get container status \"1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27\": rpc error: code = NotFound desc = could not find container \"1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27\": container with ID starting with 1675f985d615308b5c4a0b4fd748aaad60b4877e5f1f873c43b11ecb16ec2a27 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.441828 4758 scope.go:117] "RemoveContainer" containerID="fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.442461 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422\": container with ID starting with fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422 not found: ID does not exist" containerID="fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.442518 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422"} err="failed to get container status \"fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422\": rpc error: code = NotFound desc = could not find container \"fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422\": container with ID starting with fdb72a5947aff776483c38290855aaf1b563f827ce8e48419119f3fdd3933422 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.442554 4758 scope.go:117] "RemoveContainer" containerID="1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.443049 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e\": container with ID starting with 1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e not found: ID does not exist" containerID="1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.443078 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e"} err="failed to get container status \"1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e\": rpc error: code = NotFound desc = could not find container \"1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e\": container with ID starting with 1ae0322bde884f57f7a9d086420a136be58618ff1bd6c184e13709bcdfd94f2e not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.443096 4758 scope.go:117] "RemoveContainer" containerID="1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.443942 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f\": container with ID starting with 1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f not found: ID does not exist" containerID="1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.443990 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f"} err="failed to get container status \"1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f\": rpc error: code = NotFound desc = could not find container \"1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f\": container with ID starting with 1be6528bc0d5930a70e490a05239fad0858b3a0f7a7efd16d614cf924b23d01f not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.444016 4758 scope.go:117] "RemoveContainer" containerID="187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.444392 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637\": container with ID starting with 187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637 not found: ID does not exist" containerID="187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.444425 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637"} err="failed to get container status \"187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637\": rpc error: code = NotFound desc = could not find container \"187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637\": container with ID starting with 187362d8e4ec062754169e2f87ec94f7d5793fa711fa5fd15c410f1fc28bd637 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.444441 4758 scope.go:117] "RemoveContainer" containerID="98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.444851 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06\": container with ID starting with 98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06 not found: ID does not exist" containerID="98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.444881 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06"} err="failed to get container status \"98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06\": rpc error: code = NotFound desc = could not find container \"98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06\": container with ID starting with 98dba96f3d82d2f91ac1549913d63dad9a05e289a3aa4e4496710f68b5427c06 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.444902 4758 scope.go:117] "RemoveContainer" containerID="5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.445273 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81\": container with ID starting with 5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81 not found: ID does not exist" containerID="5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.445304 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81"} err="failed to get container status \"5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81\": rpc error: code = NotFound desc = could not find container \"5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81\": container with ID starting with 5d66202d2a5e17503b657fb802db140a15455bd8c4194a21b6d9f1b64bd0be81 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.445320 4758 scope.go:117] "RemoveContainer" containerID="a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.445605 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150\": container with ID starting with a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150 not found: ID does not exist" containerID="a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.445628 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150"} err="failed to get container status \"a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150\": rpc error: code = NotFound desc = could not find container \"a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150\": container with ID starting with a61d505be49b94c79eb8b827ec03e31e326e071c3582d35c774517c13fad9150 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.445644 4758 scope.go:117] "RemoveContainer" containerID="4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.446093 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b\": container with ID starting with 4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b not found: ID does not exist" containerID="4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.446121 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b"} err="failed to get container status \"4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b\": rpc error: code = NotFound desc = could not find container \"4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b\": container with ID starting with 4ec8dbd126d6c7b6939058d8b62f98cbbb570d08c9f19a41ed2f41167111a25b not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.446140 4758 scope.go:117] "RemoveContainer" containerID="48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.446473 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d\": container with ID starting with 48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d not found: ID does not exist" containerID="48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.446499 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d"} err="failed to get container status \"48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d\": rpc error: code = NotFound desc = could not find container \"48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d\": container with ID starting with 48960133cae62f67bd42247219f40d464730580e6231a3b9c3071eb58b0db66d not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.446519 4758 scope.go:117] "RemoveContainer" containerID="dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.446857 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025\": container with ID starting with dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025 not found: ID does not exist" containerID="dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.446896 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025"} err="failed to get container status \"dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025\": rpc error: code = NotFound desc = could not find container \"dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025\": container with ID starting with dc7112cf05f866ef10590fb79f557bc8c7cdec35b10225afe4f70176adc66025 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.446918 4758 scope.go:117] "RemoveContainer" containerID="7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.447288 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193\": container with ID starting with 7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193 not found: ID does not exist" containerID="7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.447315 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193"} err="failed to get container status \"7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193\": rpc error: code = NotFound desc = could not find container \"7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193\": container with ID starting with 7fdbf6254171c62c406abfd7a5157e5e9ca75689d214377da0a8becaeb120193 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.447333 4758 scope.go:117] "RemoveContainer" containerID="8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.447702 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb\": container with ID starting with 8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb not found: ID does not exist" containerID="8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.447729 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb"} err="failed to get container status \"8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb\": rpc error: code = NotFound desc = could not find container \"8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb\": container with ID starting with 8eeb45f004dbf3c1c84285eb8aaa26d7a07a1e476f1754f3cfbed1c68141b9bb not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.447748 4758 scope.go:117] "RemoveContainer" containerID="93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.448162 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75\": container with ID starting with 93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75 not found: ID does not exist" containerID="93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.448184 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75"} err="failed to get container status \"93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75\": rpc error: code = NotFound desc = could not find container \"93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75\": container with ID starting with 93b452978fc7f3f45b9681ebfd1b0470937da1e25cdca6c66092a851c0c3fb75 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.448199 4758 scope.go:117] "RemoveContainer" containerID="def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.448504 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19\": container with ID starting with def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19 not found: ID does not exist" containerID="def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.448543 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19"} err="failed to get container status \"def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19\": rpc error: code = NotFound desc = could not find container \"def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19\": container with ID starting with def211d7f1349cafa58bdb72aac53d7ef3acc3ae3fdee5075f6543b94be9bc19 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.448565 4758 scope.go:117] "RemoveContainer" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.475086 4758 scope.go:117] "RemoveContainer" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.497800 4758 scope.go:117] "RemoveContainer" containerID="883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.534146 4758 scope.go:117] "RemoveContainer" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.534661 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70\": container with ID starting with e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70 not found: ID does not exist" containerID="e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.534742 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70"} err="failed to get container status \"e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70\": rpc error: code = NotFound desc = could not find container \"e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70\": container with ID starting with e73197f17e8e7b152b5ca4ace45957f989a27fabbcf318790bcb7a85a10a4e70 not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.534775 4758 scope.go:117] "RemoveContainer" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.535395 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f\": container with ID starting with ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f not found: ID does not exist" containerID="ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.535431 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f"} err="failed to get container status \"ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f\": rpc error: code = NotFound desc = could not find container \"ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f\": container with ID starting with ce889ebaccaf8a9f47148d84cd70bcbeb2d4c092fb2a6d63e00d0355cc27ca8f not found: ID does not exist" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.535461 4758 scope.go:117] "RemoveContainer" containerID="883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74" Dec 03 17:23:36 crc kubenswrapper[4758]: E1203 17:23:36.535797 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74\": container with ID starting with 883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74 not found: ID does not exist" containerID="883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74" Dec 03 17:23:36 crc kubenswrapper[4758]: I1203 17:23:36.535823 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74"} err="failed to get container status \"883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74\": rpc error: code = NotFound desc = could not find container \"883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74\": container with ID starting with 883159d5609f417d4203def936736b8d1685e8a6264b948fcff24f307ed24c74 not found: ID does not exist" Dec 03 17:23:37 crc kubenswrapper[4758]: I1203 17:23:37.127602 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" path="/var/lib/kubelet/pods/3e30b3f9-d35f-43a0-9f06-25ebf09ccb34/volumes" Dec 03 17:23:37 crc kubenswrapper[4758]: I1203 17:23:37.129655 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" path="/var/lib/kubelet/pods/946a9b5f-7abf-4b3e-bc51-08a45e73ff30/volumes" Dec 03 17:23:41 crc kubenswrapper[4758]: I1203 17:23:41.286995 4758 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podd50c7a0d-80f6-4c8c-a516-29dbe3a6397b"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podd50c7a0d-80f6-4c8c-a516-29dbe3a6397b] : Timed out while waiting for systemd to remove kubepods-besteffort-podd50c7a0d_80f6_4c8c_a516_29dbe3a6397b.slice" Dec 03 17:23:41 crc kubenswrapper[4758]: E1203 17:23:41.287435 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podd50c7a0d-80f6-4c8c-a516-29dbe3a6397b] : unable to destroy cgroup paths for cgroup [kubepods besteffort podd50c7a0d-80f6-4c8c-a516-29dbe3a6397b] : Timed out while waiting for systemd to remove kubepods-besteffort-podd50c7a0d_80f6_4c8c_a516_29dbe3a6397b.slice" pod="openstack/cinder-api-0" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" Dec 03 17:23:42 crc kubenswrapper[4758]: I1203 17:23:42.100327 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:23:42 crc kubenswrapper[4758]: I1203 17:23:42.127027 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:23:42 crc kubenswrapper[4758]: I1203 17:23:42.133072 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:23:43 crc kubenswrapper[4758]: I1203 17:23:43.114290 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.114999 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:23:43 crc kubenswrapper[4758]: I1203 17:23:43.124012 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" path="/var/lib/kubelet/pods/d50c7a0d-80f6-4c8c-a516-29dbe3a6397b/volumes" Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.135904 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136018 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136032 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136031 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts podName:cb717c57-e263-449f-896f-38707d61e72c nodeName:}" failed. No retries permitted until 2025-12-03 17:24:15.135996957 +0000 UTC m=+1710.337373828 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts") pod "barbicane6b7-account-delete-l2xbs" (UID: "cb717c57-e263-449f-896f-38707d61e72c") : configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136092 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts podName:fef3d86b-78d4-4b18-9a05-1c2f435f03ed nodeName:}" failed. No retries permitted until 2025-12-03 17:24:15.136078919 +0000 UTC m=+1710.337455780 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts") pod "novaapic90a-account-delete-gg2x2" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed") : configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136149 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts podName:8a9b6914-42e5-4a9c-a4da-de3cee2587e2 nodeName:}" failed. No retries permitted until 2025-12-03 17:24:15.13611183 +0000 UTC m=+1710.337488891 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts") pod "novacell0700b-account-delete-snfrl" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2") : configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136193 4758 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: E1203 17:23:43.136234 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts podName:5a11fd25-44dc-4ed6-85a5-7f10fc42302a nodeName:}" failed. No retries permitted until 2025-12-03 17:24:15.136226713 +0000 UTC m=+1710.337603774 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts") pod "glancee859-account-delete-bb4lj" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a") : configmap "openstack-scripts" not found Dec 03 17:23:43 crc kubenswrapper[4758]: I1203 17:23:43.899960 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:43 crc kubenswrapper[4758]: I1203 17:23:43.952116 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2q97\" (UniqueName: \"kubernetes.io/projected/cb717c57-e263-449f-896f-38707d61e72c-kube-api-access-t2q97\") pod \"cb717c57-e263-449f-896f-38707d61e72c\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " Dec 03 17:23:43 crc kubenswrapper[4758]: I1203 17:23:43.958507 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb717c57-e263-449f-896f-38707d61e72c-kube-api-access-t2q97" (OuterVolumeSpecName: "kube-api-access-t2q97") pod "cb717c57-e263-449f-896f-38707d61e72c" (UID: "cb717c57-e263-449f-896f-38707d61e72c"). InnerVolumeSpecName "kube-api-access-t2q97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.053886 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts\") pod \"cb717c57-e263-449f-896f-38707d61e72c\" (UID: \"cb717c57-e263-449f-896f-38707d61e72c\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.054471 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2q97\" (UniqueName: \"kubernetes.io/projected/cb717c57-e263-449f-896f-38707d61e72c-kube-api-access-t2q97\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.054922 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cb717c57-e263-449f-896f-38707d61e72c" (UID: "cb717c57-e263-449f-896f-38707d61e72c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.125462 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb717c57-e263-449f-896f-38707d61e72c" containerID="3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3" exitCode=137 Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.125529 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicane6b7-account-delete-l2xbs" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.125594 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicane6b7-account-delete-l2xbs" event={"ID":"cb717c57-e263-449f-896f-38707d61e72c","Type":"ContainerDied","Data":"3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3"} Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.125774 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicane6b7-account-delete-l2xbs" event={"ID":"cb717c57-e263-449f-896f-38707d61e72c","Type":"ContainerDied","Data":"92023f680cb8eb12a4f0a2e4e3df3fe6c7304474009d9c11a7ac0023bc23b056"} Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.125888 4758 scope.go:117] "RemoveContainer" containerID="3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.130183 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a9b6914-42e5-4a9c-a4da-de3cee2587e2" containerID="0cb9a87be8ad59f9ff8dd09d281da1a8164d0ffdaca0aabbf9f6081178c0baf1" exitCode=137 Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.130238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0700b-account-delete-snfrl" event={"ID":"8a9b6914-42e5-4a9c-a4da-de3cee2587e2","Type":"ContainerDied","Data":"0cb9a87be8ad59f9ff8dd09d281da1a8164d0ffdaca0aabbf9f6081178c0baf1"} Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.156186 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb717c57-e263-449f-896f-38707d61e72c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.160982 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicane6b7-account-delete-l2xbs"] Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.162501 4758 scope.go:117] "RemoveContainer" containerID="3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3" Dec 03 17:23:44 crc kubenswrapper[4758]: E1203 17:23:44.163330 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3\": container with ID starting with 3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3 not found: ID does not exist" containerID="3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.163400 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3"} err="failed to get container status \"3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3\": rpc error: code = NotFound desc = could not find container \"3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3\": container with ID starting with 3b2cde71007273618057d33ea3b64dc807cc089ed36709a658173be974c07fa3 not found: ID does not exist" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.167289 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicane6b7-account-delete-l2xbs"] Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.303860 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.460428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts\") pod \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.460528 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7znfc\" (UniqueName: \"kubernetes.io/projected/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-kube-api-access-7znfc\") pod \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\" (UID: \"8a9b6914-42e5-4a9c-a4da-de3cee2587e2\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.461175 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a9b6914-42e5-4a9c-a4da-de3cee2587e2" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.468319 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-kube-api-access-7znfc" (OuterVolumeSpecName: "kube-api-access-7znfc") pod "8a9b6914-42e5-4a9c-a4da-de3cee2587e2" (UID: "8a9b6914-42e5-4a9c-a4da-de3cee2587e2"). InnerVolumeSpecName "kube-api-access-7znfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.491942 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.548982 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.563331 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7znfc\" (UniqueName: \"kubernetes.io/projected/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-kube-api-access-7znfc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.563379 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a9b6914-42e5-4a9c-a4da-de3cee2587e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.664336 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts\") pod \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.664496 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6skc\" (UniqueName: \"kubernetes.io/projected/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-kube-api-access-p6skc\") pod \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\" (UID: \"5a11fd25-44dc-4ed6-85a5-7f10fc42302a\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.664542 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wctjz\" (UniqueName: \"kubernetes.io/projected/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-kube-api-access-wctjz\") pod \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.664741 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts\") pod \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\" (UID: \"fef3d86b-78d4-4b18-9a05-1c2f435f03ed\") " Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.665293 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a11fd25-44dc-4ed6-85a5-7f10fc42302a" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.665811 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fef3d86b-78d4-4b18-9a05-1c2f435f03ed" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.669241 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-kube-api-access-wctjz" (OuterVolumeSpecName: "kube-api-access-wctjz") pod "fef3d86b-78d4-4b18-9a05-1c2f435f03ed" (UID: "fef3d86b-78d4-4b18-9a05-1c2f435f03ed"). InnerVolumeSpecName "kube-api-access-wctjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.669349 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-kube-api-access-p6skc" (OuterVolumeSpecName: "kube-api-access-p6skc") pod "5a11fd25-44dc-4ed6-85a5-7f10fc42302a" (UID: "5a11fd25-44dc-4ed6-85a5-7f10fc42302a"). InnerVolumeSpecName "kube-api-access-p6skc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.767193 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.767262 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6skc\" (UniqueName: \"kubernetes.io/projected/5a11fd25-44dc-4ed6-85a5-7f10fc42302a-kube-api-access-p6skc\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.767287 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wctjz\" (UniqueName: \"kubernetes.io/projected/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-kube-api-access-wctjz\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:44 crc kubenswrapper[4758]: I1203 17:23:44.767298 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef3d86b-78d4-4b18-9a05-1c2f435f03ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.125220 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb717c57-e263-449f-896f-38707d61e72c" path="/var/lib/kubelet/pods/cb717c57-e263-449f-896f-38707d61e72c/volumes" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.147076 4758 generic.go:334] "Generic (PLEG): container finished" podID="fef3d86b-78d4-4b18-9a05-1c2f435f03ed" containerID="6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899" exitCode=137 Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.147197 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic90a-account-delete-gg2x2" event={"ID":"fef3d86b-78d4-4b18-9a05-1c2f435f03ed","Type":"ContainerDied","Data":"6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899"} Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.147237 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic90a-account-delete-gg2x2" event={"ID":"fef3d86b-78d4-4b18-9a05-1c2f435f03ed","Type":"ContainerDied","Data":"79e365a4aa65e515cdb7c413f0cf1bb09a76e0051597affff3892ecd7eebd09b"} Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.147264 4758 scope.go:117] "RemoveContainer" containerID="6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.147367 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic90a-account-delete-gg2x2" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.150734 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0700b-account-delete-snfrl" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.150782 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0700b-account-delete-snfrl" event={"ID":"8a9b6914-42e5-4a9c-a4da-de3cee2587e2","Type":"ContainerDied","Data":"94a096874bc11dec2f4b20e32ad131b600b66e413d85f688b424a26f7fdf23c8"} Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.154107 4758 generic.go:334] "Generic (PLEG): container finished" podID="5a11fd25-44dc-4ed6-85a5-7f10fc42302a" containerID="d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c" exitCode=137 Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.154160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee859-account-delete-bb4lj" event={"ID":"5a11fd25-44dc-4ed6-85a5-7f10fc42302a","Type":"ContainerDied","Data":"d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c"} Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.154190 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee859-account-delete-bb4lj" event={"ID":"5a11fd25-44dc-4ed6-85a5-7f10fc42302a","Type":"ContainerDied","Data":"210592bc487c79d65a3850b60a1230bf68b4872ed384004affb1f3ec3ce2986d"} Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.154337 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancee859-account-delete-bb4lj" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.184228 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic90a-account-delete-gg2x2"] Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.191364 4758 scope.go:117] "RemoveContainer" containerID="6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899" Dec 03 17:23:45 crc kubenswrapper[4758]: E1203 17:23:45.192017 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899\": container with ID starting with 6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899 not found: ID does not exist" containerID="6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.192088 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899"} err="failed to get container status \"6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899\": rpc error: code = NotFound desc = could not find container \"6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899\": container with ID starting with 6ca96fabb9b521c884b4e96b81a262638deae049ecd32282c3b46ae3fa98c899 not found: ID does not exist" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.192126 4758 scope.go:117] "RemoveContainer" containerID="0cb9a87be8ad59f9ff8dd09d281da1a8164d0ffdaca0aabbf9f6081178c0baf1" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.197192 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapic90a-account-delete-gg2x2"] Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.208867 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0700b-account-delete-snfrl"] Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.216294 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0700b-account-delete-snfrl"] Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.216824 4758 scope.go:117] "RemoveContainer" containerID="d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.223794 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancee859-account-delete-bb4lj"] Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.231105 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancee859-account-delete-bb4lj"] Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.244312 4758 scope.go:117] "RemoveContainer" containerID="d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c" Dec 03 17:23:45 crc kubenswrapper[4758]: E1203 17:23:45.244758 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c\": container with ID starting with d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c not found: ID does not exist" containerID="d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c" Dec 03 17:23:45 crc kubenswrapper[4758]: I1203 17:23:45.244801 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c"} err="failed to get container status \"d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c\": rpc error: code = NotFound desc = could not find container \"d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c\": container with ID starting with d84bfdb875621612c2f92d1368d341e2f5eb2961b2d342c728c46ff95e4aa13c not found: ID does not exist" Dec 03 17:23:47 crc kubenswrapper[4758]: I1203 17:23:47.125291 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a11fd25-44dc-4ed6-85a5-7f10fc42302a" path="/var/lib/kubelet/pods/5a11fd25-44dc-4ed6-85a5-7f10fc42302a/volumes" Dec 03 17:23:47 crc kubenswrapper[4758]: I1203 17:23:47.126218 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a9b6914-42e5-4a9c-a4da-de3cee2587e2" path="/var/lib/kubelet/pods/8a9b6914-42e5-4a9c-a4da-de3cee2587e2/volumes" Dec 03 17:23:47 crc kubenswrapper[4758]: I1203 17:23:47.126662 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef3d86b-78d4-4b18-9a05-1c2f435f03ed" path="/var/lib/kubelet/pods/fef3d86b-78d4-4b18-9a05-1c2f435f03ed/volumes" Dec 03 17:23:58 crc kubenswrapper[4758]: I1203 17:23:58.115314 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:23:58 crc kubenswrapper[4758]: E1203 17:23:58.116205 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.471663 4758 scope.go:117] "RemoveContainer" containerID="dc93f50cb6ec038cc05b073430b2e0b33bcdb3500813955c37b1871f39af4d2d" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.509037 4758 scope.go:117] "RemoveContainer" containerID="43db26e3f86de23e8be92405726fa2e1dfd092fdd6eed2ce63132cd949d1945d" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.531958 4758 scope.go:117] "RemoveContainer" containerID="067f55e5df424ba5ecc4abf589ef0908d61bb0bb9e7faae3ba23b2e137ee2b4f" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.573415 4758 scope.go:117] "RemoveContainer" containerID="e524b6d9bd3a5f4290b198d37c9f8d00e0086aebe49103ce51921e9f8903e0c9" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.609493 4758 scope.go:117] "RemoveContainer" containerID="83f9b29f9b61e08062efbeb4039c8c3a3d37c917f7dfb5c674091aaa5101504a" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.629524 4758 scope.go:117] "RemoveContainer" containerID="23cd376c93f44deb3713feb4e1976b83962a216ee24150ad8c579fa17ac4bff5" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.676263 4758 scope.go:117] "RemoveContainer" containerID="c608fc7bb4625c50cf87800dcbdbdd20deaaadae0ba69d4b513648ae6e5178bd" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.704642 4758 scope.go:117] "RemoveContainer" containerID="1f7ba867db015c485ed3d0570727a2b99fbe07212ced2bb21d6f5f4e010efe7e" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.735793 4758 scope.go:117] "RemoveContainer" containerID="32e6077c9e48fa700c66888afcec4d226e098bafa50bf0a95011cf9ddf7f603e" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.767602 4758 scope.go:117] "RemoveContainer" containerID="a74f15a6d511a8bec6e690deb81da4dadb59f18bcc62fea8232cd90e22e04883" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.796508 4758 scope.go:117] "RemoveContainer" containerID="412987fdb9dbd3120fe67645f65ca5836f91df6e24f7be18c9b5b46a4e6cccb1" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.818886 4758 scope.go:117] "RemoveContainer" containerID="d634789b8f8c0e90fd890b15f1ea496945238b76512f4197c7fe20fa25e0eaa6" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.845357 4758 scope.go:117] "RemoveContainer" containerID="23cfb2637f3a883c7374c352adf87946c523400b25fa81403565d45084fa761d" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.881899 4758 scope.go:117] "RemoveContainer" containerID="9804e5f426b099dbf4ce321f0e6e9c4664e9582bdfe2efda45780608e8fea933" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.917818 4758 scope.go:117] "RemoveContainer" containerID="664e30dda3d80d735a4b98f34b975e32bed09a6c3a4d4ec313abc561a54aae69" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.950880 4758 scope.go:117] "RemoveContainer" containerID="d14d6da4aea8138b0af55c3d28863a692ecfb55bd249cf3282b81c936b6f5a42" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.977243 4758 scope.go:117] "RemoveContainer" containerID="e1b3122e4ed5eb97bfcaa944671ba839285f8b7c938d8f2fddbfabb7cdc55868" Dec 03 17:24:06 crc kubenswrapper[4758]: I1203 17:24:06.999802 4758 scope.go:117] "RemoveContainer" containerID="6d70a91ca70352bcd228584cfc0a35eeeed5d96455696994c096ca68be4c414f" Dec 03 17:24:13 crc kubenswrapper[4758]: I1203 17:24:13.115000 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:24:13 crc kubenswrapper[4758]: E1203 17:24:13.116110 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:24:25 crc kubenswrapper[4758]: I1203 17:24:25.118706 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:24:25 crc kubenswrapper[4758]: E1203 17:24:25.121466 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:24:37 crc kubenswrapper[4758]: I1203 17:24:37.115792 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:24:37 crc kubenswrapper[4758]: E1203 17:24:37.116877 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:24:50 crc kubenswrapper[4758]: I1203 17:24:50.115728 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:24:50 crc kubenswrapper[4758]: E1203 17:24:50.117101 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:25:04 crc kubenswrapper[4758]: I1203 17:25:04.114535 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:25:04 crc kubenswrapper[4758]: E1203 17:25:04.115422 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.785514 4758 scope.go:117] "RemoveContainer" containerID="7b60c04c03cbdf8b0711c20b25062dc690eb658b38baf1bf3b829e0decd6d56d" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.818134 4758 scope.go:117] "RemoveContainer" containerID="7247d58e71f563549d53c42917cec0911f2f98ae88698a1caa308986151490b6" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.879712 4758 scope.go:117] "RemoveContainer" containerID="25065c61a15ddba760fbd565f929949c6cd3b972d95f83ea74bea465a8ba22dc" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.906238 4758 scope.go:117] "RemoveContainer" containerID="288ec7994c38909faa37c6fb97e7c615461c154f58fffb1db5fac5164c538a3c" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.947151 4758 scope.go:117] "RemoveContainer" containerID="753c1f65ee759c0fdb55ace0c17c7d87b84db76d9c268e39c820bbac1853d21b" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.965149 4758 scope.go:117] "RemoveContainer" containerID="e28be1b4932b60511077a9023fc129e0953206f967f40a629526334ce7f94d00" Dec 03 17:25:07 crc kubenswrapper[4758]: I1203 17:25:07.987444 4758 scope.go:117] "RemoveContainer" containerID="c4f33dad18cd24c6fb9262563f5df5def2e2efa7c561f64389ce9ac18f00ff1c" Dec 03 17:25:08 crc kubenswrapper[4758]: I1203 17:25:08.023704 4758 scope.go:117] "RemoveContainer" containerID="986d8b4d0b76748ce94c47a15f1571bda7e27fd288e0506f7528909dffe76a1b" Dec 03 17:25:08 crc kubenswrapper[4758]: I1203 17:25:08.059552 4758 scope.go:117] "RemoveContainer" containerID="6c1f749418d08e981d41d7fa033a9510b208034ee61c5bce425ac2db39489187" Dec 03 17:25:08 crc kubenswrapper[4758]: I1203 17:25:08.092086 4758 scope.go:117] "RemoveContainer" containerID="cc234173c0c1f85c77be92341554d1a75f3b1857783d167b5ab17235b9336237" Dec 03 17:25:08 crc kubenswrapper[4758]: I1203 17:25:08.134397 4758 scope.go:117] "RemoveContainer" containerID="7d1a2b7de5277b2da6cd394b4671007ed245d4da0dcaddd9b381996e33f4f2e5" Dec 03 17:25:08 crc kubenswrapper[4758]: I1203 17:25:08.188872 4758 scope.go:117] "RemoveContainer" containerID="50b6eae62c328e1085de8a54db7c2d8a55349a548197be9d1d11098b72f55ec6" Dec 03 17:25:16 crc kubenswrapper[4758]: I1203 17:25:16.114896 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:25:16 crc kubenswrapper[4758]: E1203 17:25:16.116045 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:25:27 crc kubenswrapper[4758]: I1203 17:25:27.115174 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:25:27 crc kubenswrapper[4758]: E1203 17:25:27.116169 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:25:39 crc kubenswrapper[4758]: I1203 17:25:39.115072 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:25:39 crc kubenswrapper[4758]: E1203 17:25:39.115821 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:25:52 crc kubenswrapper[4758]: I1203 17:25:52.115537 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:25:52 crc kubenswrapper[4758]: E1203 17:25:52.116625 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:26:06 crc kubenswrapper[4758]: I1203 17:26:06.114811 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:26:06 crc kubenswrapper[4758]: E1203 17:26:06.115756 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.345155 4758 scope.go:117] "RemoveContainer" containerID="19b56edda648ad38df5de5fb1958faf6c38b399f4a2cdaea93793667dcf943e2" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.367132 4758 scope.go:117] "RemoveContainer" containerID="99106e146935a884c8eae3a30d8b26d2549f367f4b50ca0f4536926fc866b59a" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.386024 4758 scope.go:117] "RemoveContainer" containerID="d747aff75568811a8371fabd136f0f4d2fe3a3d7d5d4dd1c649ff970432797c5" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.406611 4758 scope.go:117] "RemoveContainer" containerID="ae9f6c0cce128861e191ebc593d6fa56a7d7fd6eeadf7eaf6a4c003584c87ce7" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.454483 4758 scope.go:117] "RemoveContainer" containerID="f2582536bf237b4937da05dde1ec781a6082f8afc2d3f323fba8ffc0d87b12ca" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.477165 4758 scope.go:117] "RemoveContainer" containerID="e5480ee329143645fdfee867011988a34e8d68ff7a64b40999efdb3782b3f44d" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.505627 4758 scope.go:117] "RemoveContainer" containerID="2bad2b11cca0d4a475d1f157b37f370905f6d7ad77235abfbb17cdd14e36cff6" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.533329 4758 scope.go:117] "RemoveContainer" containerID="c9a053a98293b8d5dd7c40e6e5514af1ae1e974be1223ffac53d82fad7efd74a" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.556466 4758 scope.go:117] "RemoveContainer" containerID="7277037e4a4d38c2ea41fe5cf0960dceaeccfafcabb4c1f19da9780663d3a688" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.583905 4758 scope.go:117] "RemoveContainer" containerID="b9a6a87f277fcfa9d46702f8fb08b9dd826f5798578c0b34f74e4508f4c77b21" Dec 03 17:26:08 crc kubenswrapper[4758]: I1203 17:26:08.608268 4758 scope.go:117] "RemoveContainer" containerID="c1e072acd4c62b59d1cb154ec06ad176f5ed4d1bdd3cc652119c966d7118b7f9" Dec 03 17:26:19 crc kubenswrapper[4758]: I1203 17:26:19.114279 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:26:19 crc kubenswrapper[4758]: E1203 17:26:19.115267 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:26:31 crc kubenswrapper[4758]: I1203 17:26:31.114734 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:26:31 crc kubenswrapper[4758]: E1203 17:26:31.115786 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:26:46 crc kubenswrapper[4758]: I1203 17:26:46.114796 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:26:46 crc kubenswrapper[4758]: E1203 17:26:46.115972 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:26:58 crc kubenswrapper[4758]: I1203 17:26:58.114878 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:26:58 crc kubenswrapper[4758]: E1203 17:26:58.115751 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:27:08 crc kubenswrapper[4758]: I1203 17:27:08.743172 4758 scope.go:117] "RemoveContainer" containerID="8a484e9fea768918ed84dcc3fd4d0716456f1c33003094d4dfccc9c14d6684a7" Dec 03 17:27:08 crc kubenswrapper[4758]: I1203 17:27:08.787792 4758 scope.go:117] "RemoveContainer" containerID="b226eaf6e02f5bf23bb058d2357c7eb31da688995dc12dbdc9fe34fe8dbc7091" Dec 03 17:27:08 crc kubenswrapper[4758]: I1203 17:27:08.812655 4758 scope.go:117] "RemoveContainer" containerID="aacdb99dee5e45686e33def0624769051138394c943550e16068c2601899b7a4" Dec 03 17:27:11 crc kubenswrapper[4758]: I1203 17:27:11.115106 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:27:11 crc kubenswrapper[4758]: E1203 17:27:11.115833 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:27:26 crc kubenswrapper[4758]: I1203 17:27:26.114937 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:27:26 crc kubenswrapper[4758]: E1203 17:27:26.118368 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:27:37 crc kubenswrapper[4758]: I1203 17:27:37.114297 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:27:37 crc kubenswrapper[4758]: E1203 17:27:37.115310 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:27:49 crc kubenswrapper[4758]: I1203 17:27:49.114918 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:27:50 crc kubenswrapper[4758]: I1203 17:27:50.416498 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"7bb067f35d75463b1a551aebb285d386625b1b5d3cad4de5ba7441f9f3fdcfc4"} Dec 03 17:28:08 crc kubenswrapper[4758]: I1203 17:28:08.870891 4758 scope.go:117] "RemoveContainer" containerID="6bf75a7b95cf3a4d7d3c7984e0780c10a5e1cb76fbe0816ad9b844b0ed2b0584" Dec 03 17:28:08 crc kubenswrapper[4758]: I1203 17:28:08.913327 4758 scope.go:117] "RemoveContainer" containerID="f8462cdf53aa4a2f1c9cf93174c1258d17adcfbc127e072e1de997ba0f6f330d" Dec 03 17:28:08 crc kubenswrapper[4758]: I1203 17:28:08.935633 4758 scope.go:117] "RemoveContainer" containerID="28190b32c88efa920b93be379a781b27859fc7ef01a1ee77d97810a89f3906f3" Dec 03 17:28:08 crc kubenswrapper[4758]: I1203 17:28:08.992699 4758 scope.go:117] "RemoveContainer" containerID="2f0d4efaf8fb3419ad631f420a3c8d2ab75387d893dc5b9e0ec89e0193a7389c" Dec 03 17:28:09 crc kubenswrapper[4758]: I1203 17:28:09.046004 4758 scope.go:117] "RemoveContainer" containerID="31c58b29312dd48a9c7ea2e97602a4641864e6411857864b3871e30db50be257" Dec 03 17:28:09 crc kubenswrapper[4758]: I1203 17:28:09.088083 4758 scope.go:117] "RemoveContainer" containerID="725fb527456d51890a2e88aeb9e163111cb380a48930dbc8ce6dd902e06c3709" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.176138 4758 scope.go:117] "RemoveContainer" containerID="08867454fb2022e6047c48092dc940ce2f60d0e3bb28b9647d08e3cc414e048e" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.206066 4758 scope.go:117] "RemoveContainer" containerID="d0fc34ea02fb5712f5c26717176864428bc2e389b2d8605da62e87ecf5f8f023" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.224619 4758 scope.go:117] "RemoveContainer" containerID="0bd8d0e46e216ea4fb96179a0303560e480f37a479d27051d4e0b9648d0aaa54" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.250740 4758 scope.go:117] "RemoveContainer" containerID="39dfeec538bbe7eb43d14ac4885d21560e65527c9acf6664be573a26763ad32b" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.272409 4758 scope.go:117] "RemoveContainer" containerID="484b8a0e84e8e29e4a83fcb1f26cd7fc46d4f41e4674efa1174c310f89876926" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.294208 4758 scope.go:117] "RemoveContainer" containerID="421b471351888521cc265529cdcf5204bbfd978298865928a87d0b8d3f8072bd" Dec 03 17:29:09 crc kubenswrapper[4758]: I1203 17:29:09.315432 4758 scope.go:117] "RemoveContainer" containerID="e1f514beb4453b1687bf12d6068c37dc748aad9f2f74d218b6b3104db6148ce3" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.312418 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jlkgc"] Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313461 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313475 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-api" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313491 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313498 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313505 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313511 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313521 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef3d86b-78d4-4b18-9a05-1c2f435f03ed" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313527 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef3d86b-78d4-4b18-9a05-1c2f435f03ed" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313541 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2923bcc5-72ec-493c-ad10-9f171801ae4a" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313547 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2923bcc5-72ec-493c-ad10-9f171801ae4a" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313557 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb717c57-e263-449f-896f-38707d61e72c" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313562 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb717c57-e263-449f-896f-38707d61e72c" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313579 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="mysql-bootstrap" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313586 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="mysql-bootstrap" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313594 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server-init" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313601 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server-init" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313610 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-expirer" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313616 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-expirer" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313624 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="galera" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313629 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="galera" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313638 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313644 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313653 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313659 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313665 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde9ec06-c65f-45a5-9e90-5a12397f010e" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313671 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde9ec06-c65f-45a5-9e90-5a12397f010e" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313695 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313701 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-server" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313714 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4e949a-2e3c-49ff-ba65-f925a9a95336" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313720 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4e949a-2e3c-49ff-ba65-f925a9a95336" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313730 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerName="rabbitmq" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313736 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerName="rabbitmq" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313747 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313753 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313760 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313766 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-server" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313776 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-updater" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313782 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-updater" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313789 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-notification-agent" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313795 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-notification-agent" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313806 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a1c0c2-3a27-43cc-8d41-d3c7508ab441" containerName="kube-state-metrics" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313811 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a1c0c2-3a27-43cc-8d41-d3c7508ab441" containerName="kube-state-metrics" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313821 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" containerName="keystone-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313827 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" containerName="keystone-api" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313837 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-metadata" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313844 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-metadata" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313854 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313860 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313866 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313872 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313879 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313885 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313896 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313901 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313909 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313916 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313924 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313930 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313941 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313948 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-api" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313957 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="sg-core" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313963 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="sg-core" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313974 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="swift-recon-cron" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313981 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="swift-recon-cron" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.313988 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.313994 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314001 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314008 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314017 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314022 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314030 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314037 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314043 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314049 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314056 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="cinder-scheduler" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314063 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="cinder-scheduler" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314069 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314075 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314083 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="ovsdbserver-sb" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314089 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="ovsdbserver-sb" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314100 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerName="init" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314106 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerName="init" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314112 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a9b6914-42e5-4a9c-a4da-de3cee2587e2" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314119 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a9b6914-42e5-4a9c-a4da-de3cee2587e2" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314127 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerName="setup-container" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314133 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerName="setup-container" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314144 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-central-agent" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314150 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-central-agent" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314156 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerName="setup-container" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314162 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerName="setup-container" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314169 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314176 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314182 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314188 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314195 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerName="nova-cell1-conductor-conductor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314201 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerName="nova-cell1-conductor-conductor" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314212 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a11fd25-44dc-4ed6-85a5-7f10fc42302a" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314217 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a11fd25-44dc-4ed6-85a5-7f10fc42302a" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314228 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebee961b-48e5-4983-bb45-c14bbcc164f8" containerName="nova-cell0-conductor-conductor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314233 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebee961b-48e5-4983-bb45-c14bbcc164f8" containerName="nova-cell0-conductor-conductor" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314241 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014153c0-72a5-46ae-9065-ff34c290745a" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314247 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="014153c0-72a5-46ae-9065-ff34c290745a" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314257 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" containerName="nova-scheduler-scheduler" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314263 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" containerName="nova-scheduler-scheduler" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314272 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314278 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314285 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314290 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314300 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314306 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-api" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314316 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314322 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314331 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-reaper" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314337 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-reaper" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314344 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314349 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314358 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="ovsdbserver-nb" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314364 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="ovsdbserver-nb" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314374 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314379 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314389 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerName="dnsmasq-dns" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314395 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerName="dnsmasq-dns" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314402 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="probe" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314408 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="probe" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314414 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314420 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314431 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="galera" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314437 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="galera" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314446 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314454 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314462 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314469 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-server" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314478 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="mysql-bootstrap" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314484 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="mysql-bootstrap" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314491 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314496 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314501 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-updater" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314508 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-updater" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314517 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="rsync" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314522 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="rsync" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314531 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314536 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314544 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerName="rabbitmq" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314550 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerName="rabbitmq" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314557 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314562 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314567 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="proxy-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314573 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="proxy-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314579 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314586 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-server" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314593 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314598 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-log" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314605 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0754ea69-5007-4e51-bd22-0efd13954199" containerName="memcached" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314610 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0754ea69-5007-4e51-bd22-0efd13954199" containerName="memcached" Dec 03 17:29:29 crc kubenswrapper[4758]: E1203 17:29:29.314618 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df131e1-f13f-4943-95a0-18086525f883" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314624 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df131e1-f13f-4943-95a0-18086525f883" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314789 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb717c57-e263-449f-896f-38707d61e72c" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314802 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314814 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="ovsdbserver-nb" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314822 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314833 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-updater" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314841 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e518390e-b4c4-49bb-913f-ddfc331456ab" containerName="rabbitmq" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314847 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314856 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="cinder-scheduler" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314863 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-notification-agent" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314870 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b3f0d2-7b01-48a6-b2be-9e205027d1a2" containerName="nova-scheduler-scheduler" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314878 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-reaper" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314887 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314895 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-updater" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314906 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314914 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314922 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a11fd25-44dc-4ed6-85a5-7f10fc42302a" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314929 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df131e1-f13f-4943-95a0-18086525f883" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314938 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="swift-recon-cron" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314946 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314956 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314965 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314972 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a1c0c2-3a27-43cc-8d41-d3c7508ab441" containerName="kube-state-metrics" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314979 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="ovsdbserver-sb" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314988 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d870c33d-67ba-4991-8624-4d0a28cb8542" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.314998 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="218264a1-c6a9-4ecb-8f40-5df1fd7dc77c" containerName="barbican-worker-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315006 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b1161a-3f70-4956-b13a-1ef693aa2e7c" containerName="galera" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315014 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b4e949a-2e3c-49ff-ba65-f925a9a95336" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315021 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf64139-0696-484a-870f-023898fbeb08" containerName="ovn-northd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315031 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-expirer" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315042 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315048 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0cc9a7c-fe92-4cfe-b8f6-a523db38b665" containerName="barbican-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315058 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315065 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="091fea73-c280-4583-810c-1cfe5f1f0fe1" containerName="dnsmasq-dns" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315071 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315081 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb5db84-3c14-46e1-a18b-2a0ac35ea34b" containerName="nova-cell1-conductor-conductor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315106 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315116 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovs-vswitchd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315122 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde9ec06-c65f-45a5-9e90-5a12397f010e" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315130 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315139 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a9b6914-42e5-4a9c-a4da-de3cee2587e2" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315145 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e30b3f9-d35f-43a0-9f06-25ebf09ccb34" containerName="ovsdb-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315151 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315158 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="rsync" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315165 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315174 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb8eeb7-c8f9-4324-a518-b837c3ad043c" containerName="probe" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315183 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="583229ad-ae2e-41b5-aad3-3c196b829f40" containerName="proxy-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315191 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="070406cc-6d65-4fd0-81d7-264adead6bbd" containerName="rabbitmq" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315197 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bdefd9e-77f0-4137-8b30-35aca9f17c3f" containerName="keystone-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315203 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b74dc9-5836-4c5c-bae8-803f64c9de12" containerName="barbican-keystone-listener" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315214 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b461991e-58f6-4ff3-81d3-9cf2288d9b65" containerName="nova-api-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315222 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac23ac3-72ea-49a7-9ded-6fabe28ff690" containerName="ovn-controller" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315232 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef3d86b-78d4-4b18-9a05-1c2f435f03ed" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315241 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c80de0e-2bee-4e3f-8edf-f4a235b01d76" containerName="galera" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315247 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2923bcc5-72ec-493c-ad10-9f171801ae4a" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315255 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="ceilometer-central-agent" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315263 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d50c7a0d-80f6-4c8c-a516-29dbe3a6397b" containerName="cinder-api" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315270 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="be629420-0e78-44d4-9ca1-4ee1601d4ca6" containerName="glance-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315279 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f328daf-ac93-432b-8645-14b7030a24d1" containerName="placement-log" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315290 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="84469de1-c706-4442-8610-31188889dad1" containerName="nova-metadata-metadata" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315300 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b49b1d3-cbb4-49da-8244-69e15595da33" containerName="glance-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315309 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="sg-core" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315315 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="014153c0-72a5-46ae-9065-ff34c290745a" containerName="mariadb-account-delete" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315323 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd709657-fb8d-4c81-873f-a0b44d63abef" containerName="neutron-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315331 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebee961b-48e5-4983-bb45-c14bbcc164f8" containerName="nova-cell0-conductor-conductor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315338 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="container-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315346 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb44365-2dda-4769-813a-712f362fff6b" containerName="proxy-httpd" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315355 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-auditor" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315363 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0754ea69-5007-4e51-bd22-0efd13954199" containerName="memcached" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315371 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a370910-1957-4bd8-891b-7f1065c5e286" containerName="openstack-network-exporter" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315379 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315389 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="account-server" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.315398 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="946a9b5f-7abf-4b3e-bc51-08a45e73ff30" containerName="object-replicator" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.320721 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.330431 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlkgc"] Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.418347 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-catalog-content\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.418449 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n4hf\" (UniqueName: \"kubernetes.io/projected/1c16287e-19de-4493-bef9-f264c31e8be4-kube-api-access-2n4hf\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.418534 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-utilities\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.520652 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-utilities\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.520870 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-catalog-content\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.520931 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n4hf\" (UniqueName: \"kubernetes.io/projected/1c16287e-19de-4493-bef9-f264c31e8be4-kube-api-access-2n4hf\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.521316 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-utilities\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.521598 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-catalog-content\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.542879 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n4hf\" (UniqueName: \"kubernetes.io/projected/1c16287e-19de-4493-bef9-f264c31e8be4-kube-api-access-2n4hf\") pod \"certified-operators-jlkgc\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.647993 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.915696 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cc94j"] Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.918197 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:29 crc kubenswrapper[4758]: I1203 17:29:29.943324 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cc94j"] Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.038994 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-utilities\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.039086 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmxdp\" (UniqueName: \"kubernetes.io/projected/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-kube-api-access-wmxdp\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.039113 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-catalog-content\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.143221 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-utilities\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.143292 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmxdp\" (UniqueName: \"kubernetes.io/projected/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-kube-api-access-wmxdp\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.143322 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-catalog-content\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.143976 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-catalog-content\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.144310 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-utilities\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.169177 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmxdp\" (UniqueName: \"kubernetes.io/projected/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-kube-api-access-wmxdp\") pod \"redhat-marketplace-cc94j\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.191421 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlkgc"] Dec 03 17:29:30 crc kubenswrapper[4758]: W1203 17:29:30.201271 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c16287e_19de_4493_bef9_f264c31e8be4.slice/crio-fda1a1466f74809ef8cb019d075f23238b5fb9ede3237cc716956a2a3c1babca WatchSource:0}: Error finding container fda1a1466f74809ef8cb019d075f23238b5fb9ede3237cc716956a2a3c1babca: Status 404 returned error can't find the container with id fda1a1466f74809ef8cb019d075f23238b5fb9ede3237cc716956a2a3c1babca Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.261749 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.271547 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerStarted","Data":"fda1a1466f74809ef8cb019d075f23238b5fb9ede3237cc716956a2a3c1babca"} Dec 03 17:29:30 crc kubenswrapper[4758]: I1203 17:29:30.717796 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cc94j"] Dec 03 17:29:30 crc kubenswrapper[4758]: W1203 17:29:30.726876 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2e14fc0_91cb_4d30_8221_30c1fa3b70f1.slice/crio-39620f06c0e1d09dd4f4ddf35da97e64cd50a4fc7fbabfb25cc877020f5a4a6a WatchSource:0}: Error finding container 39620f06c0e1d09dd4f4ddf35da97e64cd50a4fc7fbabfb25cc877020f5a4a6a: Status 404 returned error can't find the container with id 39620f06c0e1d09dd4f4ddf35da97e64cd50a4fc7fbabfb25cc877020f5a4a6a Dec 03 17:29:31 crc kubenswrapper[4758]: I1203 17:29:31.280925 4758 generic.go:334] "Generic (PLEG): container finished" podID="1c16287e-19de-4493-bef9-f264c31e8be4" containerID="0b9eb5a79513f6f11b28c70ac5f92f036744b2a06790056c60f9340166b00663" exitCode=0 Dec 03 17:29:31 crc kubenswrapper[4758]: I1203 17:29:31.281035 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerDied","Data":"0b9eb5a79513f6f11b28c70ac5f92f036744b2a06790056c60f9340166b00663"} Dec 03 17:29:31 crc kubenswrapper[4758]: I1203 17:29:31.283501 4758 generic.go:334] "Generic (PLEG): container finished" podID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerID="d3324336ff9989999a9b5685c957f861d55d1847f871bb7bb20006defc96da63" exitCode=0 Dec 03 17:29:31 crc kubenswrapper[4758]: I1203 17:29:31.283563 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cc94j" event={"ID":"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1","Type":"ContainerDied","Data":"d3324336ff9989999a9b5685c957f861d55d1847f871bb7bb20006defc96da63"} Dec 03 17:29:31 crc kubenswrapper[4758]: I1203 17:29:31.283597 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cc94j" event={"ID":"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1","Type":"ContainerStarted","Data":"39620f06c0e1d09dd4f4ddf35da97e64cd50a4fc7fbabfb25cc877020f5a4a6a"} Dec 03 17:29:31 crc kubenswrapper[4758]: I1203 17:29:31.283612 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.298149 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerStarted","Data":"7e10c5ad46ca9bb4bb21032228be2b00341ebc84535e9aaa457d96a3c42fe03a"} Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.309864 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72bb4"] Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.310598 4758 generic.go:334] "Generic (PLEG): container finished" podID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerID="6935138324e22af8f81eb69b49f7a1833cf4e3a7bd2cfa264a3117edda75f0ce" exitCode=0 Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.311931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cc94j" event={"ID":"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1","Type":"ContainerDied","Data":"6935138324e22af8f81eb69b49f7a1833cf4e3a7bd2cfa264a3117edda75f0ce"} Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.312073 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.339130 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72bb4"] Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.487838 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-utilities\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.487979 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npcs2\" (UniqueName: \"kubernetes.io/projected/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-kube-api-access-npcs2\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.488085 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-catalog-content\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.589215 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npcs2\" (UniqueName: \"kubernetes.io/projected/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-kube-api-access-npcs2\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.589288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-catalog-content\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.589349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-utilities\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.589822 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-utilities\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.589847 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-catalog-content\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.616156 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npcs2\" (UniqueName: \"kubernetes.io/projected/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-kube-api-access-npcs2\") pod \"community-operators-72bb4\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:32 crc kubenswrapper[4758]: I1203 17:29:32.675852 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:33 crc kubenswrapper[4758]: W1203 17:29:33.138801 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18742ffb_eeeb_4fa4_aaac_aebb7885dcef.slice/crio-d2ac889a9e6b584303d4f0b705fe7fd4e07b41e1819af501b21774e0b361af84 WatchSource:0}: Error finding container d2ac889a9e6b584303d4f0b705fe7fd4e07b41e1819af501b21774e0b361af84: Status 404 returned error can't find the container with id d2ac889a9e6b584303d4f0b705fe7fd4e07b41e1819af501b21774e0b361af84 Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.139717 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72bb4"] Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.320034 4758 generic.go:334] "Generic (PLEG): container finished" podID="1c16287e-19de-4493-bef9-f264c31e8be4" containerID="7e10c5ad46ca9bb4bb21032228be2b00341ebc84535e9aaa457d96a3c42fe03a" exitCode=0 Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.320121 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerDied","Data":"7e10c5ad46ca9bb4bb21032228be2b00341ebc84535e9aaa457d96a3c42fe03a"} Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.323399 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cc94j" event={"ID":"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1","Type":"ContainerStarted","Data":"edf9b7caaffadb6756e2be77d4b228d410b960c339f3ae1f8c1b17088cdf8b31"} Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.326206 4758 generic.go:334] "Generic (PLEG): container finished" podID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerID="641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f" exitCode=0 Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.326238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72bb4" event={"ID":"18742ffb-eeeb-4fa4-aaac-aebb7885dcef","Type":"ContainerDied","Data":"641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f"} Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.326256 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72bb4" event={"ID":"18742ffb-eeeb-4fa4-aaac-aebb7885dcef","Type":"ContainerStarted","Data":"d2ac889a9e6b584303d4f0b705fe7fd4e07b41e1819af501b21774e0b361af84"} Dec 03 17:29:33 crc kubenswrapper[4758]: I1203 17:29:33.377443 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cc94j" podStartSLOduration=2.895106507 podStartE2EDuration="4.37740987s" podCreationTimestamp="2025-12-03 17:29:29 +0000 UTC" firstStartedPulling="2025-12-03 17:29:31.287740918 +0000 UTC m=+2026.489117779" lastFinishedPulling="2025-12-03 17:29:32.770044281 +0000 UTC m=+2027.971421142" observedRunningTime="2025-12-03 17:29:33.374088891 +0000 UTC m=+2028.575465782" watchObservedRunningTime="2025-12-03 17:29:33.37740987 +0000 UTC m=+2028.578786731" Dec 03 17:29:34 crc kubenswrapper[4758]: I1203 17:29:34.337381 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerStarted","Data":"2d181518604bfbee603478874016e7f44e0bfab412e5c512737a1d49e6c3061e"} Dec 03 17:29:34 crc kubenswrapper[4758]: I1203 17:29:34.359937 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jlkgc" podStartSLOduration=2.933634536 podStartE2EDuration="5.359909866s" podCreationTimestamp="2025-12-03 17:29:29 +0000 UTC" firstStartedPulling="2025-12-03 17:29:31.283263629 +0000 UTC m=+2026.484640510" lastFinishedPulling="2025-12-03 17:29:33.709538979 +0000 UTC m=+2028.910915840" observedRunningTime="2025-12-03 17:29:34.359146106 +0000 UTC m=+2029.560522967" watchObservedRunningTime="2025-12-03 17:29:34.359909866 +0000 UTC m=+2029.561286747" Dec 03 17:29:35 crc kubenswrapper[4758]: I1203 17:29:35.375153 4758 generic.go:334] "Generic (PLEG): container finished" podID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerID="7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c" exitCode=0 Dec 03 17:29:35 crc kubenswrapper[4758]: I1203 17:29:35.375293 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72bb4" event={"ID":"18742ffb-eeeb-4fa4-aaac-aebb7885dcef","Type":"ContainerDied","Data":"7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c"} Dec 03 17:29:36 crc kubenswrapper[4758]: I1203 17:29:36.384949 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72bb4" event={"ID":"18742ffb-eeeb-4fa4-aaac-aebb7885dcef","Type":"ContainerStarted","Data":"976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0"} Dec 03 17:29:36 crc kubenswrapper[4758]: I1203 17:29:36.418411 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72bb4" podStartSLOduration=2.001567488 podStartE2EDuration="4.418379825s" podCreationTimestamp="2025-12-03 17:29:32 +0000 UTC" firstStartedPulling="2025-12-03 17:29:33.328225747 +0000 UTC m=+2028.529602608" lastFinishedPulling="2025-12-03 17:29:35.745038084 +0000 UTC m=+2030.946414945" observedRunningTime="2025-12-03 17:29:36.412075106 +0000 UTC m=+2031.613451977" watchObservedRunningTime="2025-12-03 17:29:36.418379825 +0000 UTC m=+2031.619756686" Dec 03 17:29:39 crc kubenswrapper[4758]: I1203 17:29:39.648874 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:39 crc kubenswrapper[4758]: I1203 17:29:39.649326 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:39 crc kubenswrapper[4758]: I1203 17:29:39.690351 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:40 crc kubenswrapper[4758]: I1203 17:29:40.261921 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:40 crc kubenswrapper[4758]: I1203 17:29:40.262894 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:40 crc kubenswrapper[4758]: I1203 17:29:40.301371 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:40 crc kubenswrapper[4758]: I1203 17:29:40.454083 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:40 crc kubenswrapper[4758]: I1203 17:29:40.456231 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.502486 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jlkgc"] Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.502856 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jlkgc" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="registry-server" containerID="cri-o://2d181518604bfbee603478874016e7f44e0bfab412e5c512737a1d49e6c3061e" gracePeriod=2 Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.676363 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.676881 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.716174 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cc94j"] Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.716847 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cc94j" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="registry-server" containerID="cri-o://edf9b7caaffadb6756e2be77d4b228d410b960c339f3ae1f8c1b17088cdf8b31" gracePeriod=2 Dec 03 17:29:42 crc kubenswrapper[4758]: I1203 17:29:42.729297 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:43 crc kubenswrapper[4758]: I1203 17:29:43.484940 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.454876 4758 generic.go:334] "Generic (PLEG): container finished" podID="1c16287e-19de-4493-bef9-f264c31e8be4" containerID="2d181518604bfbee603478874016e7f44e0bfab412e5c512737a1d49e6c3061e" exitCode=0 Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.454973 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerDied","Data":"2d181518604bfbee603478874016e7f44e0bfab412e5c512737a1d49e6c3061e"} Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.457455 4758 generic.go:334] "Generic (PLEG): container finished" podID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerID="edf9b7caaffadb6756e2be77d4b228d410b960c339f3ae1f8c1b17088cdf8b31" exitCode=0 Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.457492 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cc94j" event={"ID":"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1","Type":"ContainerDied","Data":"edf9b7caaffadb6756e2be77d4b228d410b960c339f3ae1f8c1b17088cdf8b31"} Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.783836 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.791465 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n4hf\" (UniqueName: \"kubernetes.io/projected/1c16287e-19de-4493-bef9-f264c31e8be4-kube-api-access-2n4hf\") pod \"1c16287e-19de-4493-bef9-f264c31e8be4\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.791569 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-catalog-content\") pod \"1c16287e-19de-4493-bef9-f264c31e8be4\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.791646 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-utilities\") pod \"1c16287e-19de-4493-bef9-f264c31e8be4\" (UID: \"1c16287e-19de-4493-bef9-f264c31e8be4\") " Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.792828 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-utilities" (OuterVolumeSpecName: "utilities") pod "1c16287e-19de-4493-bef9-f264c31e8be4" (UID: "1c16287e-19de-4493-bef9-f264c31e8be4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.800018 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c16287e-19de-4493-bef9-f264c31e8be4-kube-api-access-2n4hf" (OuterVolumeSpecName: "kube-api-access-2n4hf") pod "1c16287e-19de-4493-bef9-f264c31e8be4" (UID: "1c16287e-19de-4493-bef9-f264c31e8be4"). InnerVolumeSpecName "kube-api-access-2n4hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.845514 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c16287e-19de-4493-bef9-f264c31e8be4" (UID: "1c16287e-19de-4493-bef9-f264c31e8be4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.894346 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.894393 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n4hf\" (UniqueName: \"kubernetes.io/projected/1c16287e-19de-4493-bef9-f264c31e8be4-kube-api-access-2n4hf\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.894411 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c16287e-19de-4493-bef9-f264c31e8be4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:44 crc kubenswrapper[4758]: I1203 17:29:44.902250 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72bb4"] Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.464533 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.488707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlkgc" event={"ID":"1c16287e-19de-4493-bef9-f264c31e8be4","Type":"ContainerDied","Data":"fda1a1466f74809ef8cb019d075f23238b5fb9ede3237cc716956a2a3c1babca"} Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.488782 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlkgc" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.488817 4758 scope.go:117] "RemoveContainer" containerID="2d181518604bfbee603478874016e7f44e0bfab412e5c512737a1d49e6c3061e" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.524351 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cc94j" event={"ID":"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1","Type":"ContainerDied","Data":"39620f06c0e1d09dd4f4ddf35da97e64cd50a4fc7fbabfb25cc877020f5a4a6a"} Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.524412 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-72bb4" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="registry-server" containerID="cri-o://976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0" gracePeriod=2 Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.524532 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cc94j" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.543730 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jlkgc"] Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.561730 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jlkgc"] Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.576719 4758 scope.go:117] "RemoveContainer" containerID="7e10c5ad46ca9bb4bb21032228be2b00341ebc84535e9aaa457d96a3c42fe03a" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.602367 4758 scope.go:117] "RemoveContainer" containerID="0b9eb5a79513f6f11b28c70ac5f92f036744b2a06790056c60f9340166b00663" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.609844 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-catalog-content\") pod \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.610579 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-utilities\") pod \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.610659 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmxdp\" (UniqueName: \"kubernetes.io/projected/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-kube-api-access-wmxdp\") pod \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\" (UID: \"b2e14fc0-91cb-4d30-8221-30c1fa3b70f1\") " Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.617416 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-utilities" (OuterVolumeSpecName: "utilities") pod "b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" (UID: "b2e14fc0-91cb-4d30-8221-30c1fa3b70f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.632259 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-kube-api-access-wmxdp" (OuterVolumeSpecName: "kube-api-access-wmxdp") pod "b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" (UID: "b2e14fc0-91cb-4d30-8221-30c1fa3b70f1"). InnerVolumeSpecName "kube-api-access-wmxdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.646560 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" (UID: "b2e14fc0-91cb-4d30-8221-30c1fa3b70f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.679146 4758 scope.go:117] "RemoveContainer" containerID="edf9b7caaffadb6756e2be77d4b228d410b960c339f3ae1f8c1b17088cdf8b31" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.696402 4758 scope.go:117] "RemoveContainer" containerID="6935138324e22af8f81eb69b49f7a1833cf4e3a7bd2cfa264a3117edda75f0ce" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.714080 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.714127 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.714139 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmxdp\" (UniqueName: \"kubernetes.io/projected/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1-kube-api-access-wmxdp\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.714155 4758 scope.go:117] "RemoveContainer" containerID="d3324336ff9989999a9b5685c957f861d55d1847f871bb7bb20006defc96da63" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.890023 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.905418 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cc94j"] Dec 03 17:29:45 crc kubenswrapper[4758]: I1203 17:29:45.911761 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cc94j"] Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.018388 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-catalog-content\") pod \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.018507 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npcs2\" (UniqueName: \"kubernetes.io/projected/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-kube-api-access-npcs2\") pod \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.018579 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-utilities\") pod \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\" (UID: \"18742ffb-eeeb-4fa4-aaac-aebb7885dcef\") " Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.019622 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-utilities" (OuterVolumeSpecName: "utilities") pod "18742ffb-eeeb-4fa4-aaac-aebb7885dcef" (UID: "18742ffb-eeeb-4fa4-aaac-aebb7885dcef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.024356 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-kube-api-access-npcs2" (OuterVolumeSpecName: "kube-api-access-npcs2") pod "18742ffb-eeeb-4fa4-aaac-aebb7885dcef" (UID: "18742ffb-eeeb-4fa4-aaac-aebb7885dcef"). InnerVolumeSpecName "kube-api-access-npcs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.076183 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18742ffb-eeeb-4fa4-aaac-aebb7885dcef" (UID: "18742ffb-eeeb-4fa4-aaac-aebb7885dcef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.120207 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.120237 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.120251 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npcs2\" (UniqueName: \"kubernetes.io/projected/18742ffb-eeeb-4fa4-aaac-aebb7885dcef-kube-api-access-npcs2\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.537072 4758 generic.go:334] "Generic (PLEG): container finished" podID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerID="976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0" exitCode=0 Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.537136 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72bb4" event={"ID":"18742ffb-eeeb-4fa4-aaac-aebb7885dcef","Type":"ContainerDied","Data":"976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0"} Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.537172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72bb4" event={"ID":"18742ffb-eeeb-4fa4-aaac-aebb7885dcef","Type":"ContainerDied","Data":"d2ac889a9e6b584303d4f0b705fe7fd4e07b41e1819af501b21774e0b361af84"} Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.537188 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72bb4" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.537201 4758 scope.go:117] "RemoveContainer" containerID="976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.573524 4758 scope.go:117] "RemoveContainer" containerID="7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.575217 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72bb4"] Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.581801 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-72bb4"] Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.593156 4758 scope.go:117] "RemoveContainer" containerID="641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.609869 4758 scope.go:117] "RemoveContainer" containerID="976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0" Dec 03 17:29:46 crc kubenswrapper[4758]: E1203 17:29:46.610188 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0\": container with ID starting with 976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0 not found: ID does not exist" containerID="976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.610236 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0"} err="failed to get container status \"976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0\": rpc error: code = NotFound desc = could not find container \"976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0\": container with ID starting with 976edd680139238e6a0749eff69dea018335f6affd139a6770ffb91dfac512d0 not found: ID does not exist" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.610274 4758 scope.go:117] "RemoveContainer" containerID="7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c" Dec 03 17:29:46 crc kubenswrapper[4758]: E1203 17:29:46.610593 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c\": container with ID starting with 7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c not found: ID does not exist" containerID="7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.610640 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c"} err="failed to get container status \"7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c\": rpc error: code = NotFound desc = could not find container \"7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c\": container with ID starting with 7a7e19146dbab9d52e9b39fbf635de5bd06db2f94334bd1f8f03cccdfabb988c not found: ID does not exist" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.610720 4758 scope.go:117] "RemoveContainer" containerID="641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f" Dec 03 17:29:46 crc kubenswrapper[4758]: E1203 17:29:46.611169 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f\": container with ID starting with 641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f not found: ID does not exist" containerID="641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f" Dec 03 17:29:46 crc kubenswrapper[4758]: I1203 17:29:46.611208 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f"} err="failed to get container status \"641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f\": rpc error: code = NotFound desc = could not find container \"641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f\": container with ID starting with 641255450a3573430b0b911376add83f69841bfc8af74fed1a93269796393d7f not found: ID does not exist" Dec 03 17:29:47 crc kubenswrapper[4758]: I1203 17:29:47.133546 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" path="/var/lib/kubelet/pods/18742ffb-eeeb-4fa4-aaac-aebb7885dcef/volumes" Dec 03 17:29:47 crc kubenswrapper[4758]: I1203 17:29:47.135108 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" path="/var/lib/kubelet/pods/1c16287e-19de-4493-bef9-f264c31e8be4/volumes" Dec 03 17:29:47 crc kubenswrapper[4758]: I1203 17:29:47.136399 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" path="/var/lib/kubelet/pods/b2e14fc0-91cb-4d30-8221-30c1fa3b70f1/volumes" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.151707 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6"] Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152788 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152804 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152823 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152828 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152842 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152848 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152856 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152864 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152877 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152886 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152904 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152912 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152925 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152934 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152946 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152955 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: E1203 17:30:00.152975 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.152982 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.153154 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c16287e-19de-4493-bef9-f264c31e8be4" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.153183 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="18742ffb-eeeb-4fa4-aaac-aebb7885dcef" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.153191 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2e14fc0-91cb-4d30-8221-30c1fa3b70f1" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.153936 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.157114 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.157227 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.164697 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6"] Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.265278 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee99b637-61db-4e76-8ef8-05da30ba956b-secret-volume\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.265425 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx2zg\" (UniqueName: \"kubernetes.io/projected/ee99b637-61db-4e76-8ef8-05da30ba956b-kube-api-access-kx2zg\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.265499 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee99b637-61db-4e76-8ef8-05da30ba956b-config-volume\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.367498 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee99b637-61db-4e76-8ef8-05da30ba956b-secret-volume\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.367623 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx2zg\" (UniqueName: \"kubernetes.io/projected/ee99b637-61db-4e76-8ef8-05da30ba956b-kube-api-access-kx2zg\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.367700 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee99b637-61db-4e76-8ef8-05da30ba956b-config-volume\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.368859 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee99b637-61db-4e76-8ef8-05da30ba956b-config-volume\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.379501 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee99b637-61db-4e76-8ef8-05da30ba956b-secret-volume\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.388023 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx2zg\" (UniqueName: \"kubernetes.io/projected/ee99b637-61db-4e76-8ef8-05da30ba956b-kube-api-access-kx2zg\") pod \"collect-profiles-29413050-dsbm6\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.477045 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:00 crc kubenswrapper[4758]: I1203 17:30:00.896334 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6"] Dec 03 17:30:00 crc kubenswrapper[4758]: W1203 17:30:00.899879 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee99b637_61db_4e76_8ef8_05da30ba956b.slice/crio-047a8b897ae9fc2edd3a4f9cb16f3ae64fddd204340447df96a85fc7de44bb24 WatchSource:0}: Error finding container 047a8b897ae9fc2edd3a4f9cb16f3ae64fddd204340447df96a85fc7de44bb24: Status 404 returned error can't find the container with id 047a8b897ae9fc2edd3a4f9cb16f3ae64fddd204340447df96a85fc7de44bb24 Dec 03 17:30:01 crc kubenswrapper[4758]: I1203 17:30:01.673420 4758 generic.go:334] "Generic (PLEG): container finished" podID="ee99b637-61db-4e76-8ef8-05da30ba956b" containerID="7fbb00cf742a12cdf4376c50ef72949a4d018f1ef4d49fa3ce11599f1a8f7c66" exitCode=0 Dec 03 17:30:01 crc kubenswrapper[4758]: I1203 17:30:01.673783 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" event={"ID":"ee99b637-61db-4e76-8ef8-05da30ba956b","Type":"ContainerDied","Data":"7fbb00cf742a12cdf4376c50ef72949a4d018f1ef4d49fa3ce11599f1a8f7c66"} Dec 03 17:30:01 crc kubenswrapper[4758]: I1203 17:30:01.673818 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" event={"ID":"ee99b637-61db-4e76-8ef8-05da30ba956b","Type":"ContainerStarted","Data":"047a8b897ae9fc2edd3a4f9cb16f3ae64fddd204340447df96a85fc7de44bb24"} Dec 03 17:30:02 crc kubenswrapper[4758]: I1203 17:30:02.939298 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.013440 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx2zg\" (UniqueName: \"kubernetes.io/projected/ee99b637-61db-4e76-8ef8-05da30ba956b-kube-api-access-kx2zg\") pod \"ee99b637-61db-4e76-8ef8-05da30ba956b\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.013592 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee99b637-61db-4e76-8ef8-05da30ba956b-secret-volume\") pod \"ee99b637-61db-4e76-8ef8-05da30ba956b\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.013671 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee99b637-61db-4e76-8ef8-05da30ba956b-config-volume\") pod \"ee99b637-61db-4e76-8ef8-05da30ba956b\" (UID: \"ee99b637-61db-4e76-8ef8-05da30ba956b\") " Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.015723 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee99b637-61db-4e76-8ef8-05da30ba956b-config-volume" (OuterVolumeSpecName: "config-volume") pod "ee99b637-61db-4e76-8ef8-05da30ba956b" (UID: "ee99b637-61db-4e76-8ef8-05da30ba956b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.024052 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee99b637-61db-4e76-8ef8-05da30ba956b-kube-api-access-kx2zg" (OuterVolumeSpecName: "kube-api-access-kx2zg") pod "ee99b637-61db-4e76-8ef8-05da30ba956b" (UID: "ee99b637-61db-4e76-8ef8-05da30ba956b"). InnerVolumeSpecName "kube-api-access-kx2zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.026307 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee99b637-61db-4e76-8ef8-05da30ba956b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ee99b637-61db-4e76-8ef8-05da30ba956b" (UID: "ee99b637-61db-4e76-8ef8-05da30ba956b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.115378 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee99b637-61db-4e76-8ef8-05da30ba956b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.115608 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee99b637-61db-4e76-8ef8-05da30ba956b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.115891 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx2zg\" (UniqueName: \"kubernetes.io/projected/ee99b637-61db-4e76-8ef8-05da30ba956b-kube-api-access-kx2zg\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.694001 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.693857 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6" event={"ID":"ee99b637-61db-4e76-8ef8-05da30ba956b","Type":"ContainerDied","Data":"047a8b897ae9fc2edd3a4f9cb16f3ae64fddd204340447df96a85fc7de44bb24"} Dec 03 17:30:03 crc kubenswrapper[4758]: I1203 17:30:03.694448 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="047a8b897ae9fc2edd3a4f9cb16f3ae64fddd204340447df96a85fc7de44bb24" Dec 03 17:30:04 crc kubenswrapper[4758]: I1203 17:30:04.018037 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm"] Dec 03 17:30:04 crc kubenswrapper[4758]: I1203 17:30:04.023345 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-7jdnm"] Dec 03 17:30:05 crc kubenswrapper[4758]: I1203 17:30:05.128215 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="886f030a-c717-4277-b739-393ecb2081e4" path="/var/lib/kubelet/pods/886f030a-c717-4277-b739-393ecb2081e4/volumes" Dec 03 17:30:09 crc kubenswrapper[4758]: I1203 17:30:09.423345 4758 scope.go:117] "RemoveContainer" containerID="cb5f1783808962262d7e213cf2a4d8c33df2669539a9fb25dc259fd2cc351b56" Dec 03 17:30:11 crc kubenswrapper[4758]: I1203 17:30:11.395236 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:30:11 crc kubenswrapper[4758]: I1203 17:30:11.395339 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:30:41 crc kubenswrapper[4758]: I1203 17:30:41.394570 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:30:41 crc kubenswrapper[4758]: I1203 17:30:41.395228 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:31:11 crc kubenswrapper[4758]: I1203 17:31:11.395481 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:31:11 crc kubenswrapper[4758]: I1203 17:31:11.398106 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:31:11 crc kubenswrapper[4758]: I1203 17:31:11.398337 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:31:11 crc kubenswrapper[4758]: I1203 17:31:11.399610 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bb067f35d75463b1a551aebb285d386625b1b5d3cad4de5ba7441f9f3fdcfc4"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:31:11 crc kubenswrapper[4758]: I1203 17:31:11.399956 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://7bb067f35d75463b1a551aebb285d386625b1b5d3cad4de5ba7441f9f3fdcfc4" gracePeriod=600 Dec 03 17:31:12 crc kubenswrapper[4758]: I1203 17:31:12.329191 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="7bb067f35d75463b1a551aebb285d386625b1b5d3cad4de5ba7441f9f3fdcfc4" exitCode=0 Dec 03 17:31:12 crc kubenswrapper[4758]: I1203 17:31:12.329285 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"7bb067f35d75463b1a551aebb285d386625b1b5d3cad4de5ba7441f9f3fdcfc4"} Dec 03 17:31:12 crc kubenswrapper[4758]: I1203 17:31:12.329918 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391"} Dec 03 17:31:12 crc kubenswrapper[4758]: I1203 17:31:12.329950 4758 scope.go:117] "RemoveContainer" containerID="d785f0e174f4bf2078da774a73dc95c604158b998dcf0bcbdf680269b9a6582a" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.648356 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-55zrw"] Dec 03 17:31:39 crc kubenswrapper[4758]: E1203 17:31:39.649436 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee99b637-61db-4e76-8ef8-05da30ba956b" containerName="collect-profiles" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.649452 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee99b637-61db-4e76-8ef8-05da30ba956b" containerName="collect-profiles" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.649600 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee99b637-61db-4e76-8ef8-05da30ba956b" containerName="collect-profiles" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.650855 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.666141 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-55zrw"] Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.835623 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-utilities\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.835798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-catalog-content\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.835844 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sswh\" (UniqueName: \"kubernetes.io/projected/f108482f-1e58-4c76-a5ad-4ec9628b10e2-kube-api-access-5sswh\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.937362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-catalog-content\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.937455 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sswh\" (UniqueName: \"kubernetes.io/projected/f108482f-1e58-4c76-a5ad-4ec9628b10e2-kube-api-access-5sswh\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.937505 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-utilities\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.938375 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-utilities\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.938384 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-catalog-content\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.962941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sswh\" (UniqueName: \"kubernetes.io/projected/f108482f-1e58-4c76-a5ad-4ec9628b10e2-kube-api-access-5sswh\") pod \"redhat-operators-55zrw\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:39 crc kubenswrapper[4758]: I1203 17:31:39.970604 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:40 crc kubenswrapper[4758]: I1203 17:31:40.446364 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-55zrw"] Dec 03 17:31:40 crc kubenswrapper[4758]: I1203 17:31:40.587730 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerStarted","Data":"7bed3216b320d456171af7f23e5237d039e9c21c63d692b56ca1b099bf8d6527"} Dec 03 17:31:41 crc kubenswrapper[4758]: I1203 17:31:41.600510 4758 generic.go:334] "Generic (PLEG): container finished" podID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerID="7c35186bcc8eca4fb9ecf88b778cbcbc5904831341eeb89603f8440b7290834f" exitCode=0 Dec 03 17:31:41 crc kubenswrapper[4758]: I1203 17:31:41.600611 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerDied","Data":"7c35186bcc8eca4fb9ecf88b778cbcbc5904831341eeb89603f8440b7290834f"} Dec 03 17:31:42 crc kubenswrapper[4758]: I1203 17:31:42.610378 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerStarted","Data":"1c9e9ea0db629f50b912737a32dfacb81a4dd48835fd3cc427805539374d0537"} Dec 03 17:31:43 crc kubenswrapper[4758]: I1203 17:31:43.619761 4758 generic.go:334] "Generic (PLEG): container finished" podID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerID="1c9e9ea0db629f50b912737a32dfacb81a4dd48835fd3cc427805539374d0537" exitCode=0 Dec 03 17:31:43 crc kubenswrapper[4758]: I1203 17:31:43.619885 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerDied","Data":"1c9e9ea0db629f50b912737a32dfacb81a4dd48835fd3cc427805539374d0537"} Dec 03 17:31:44 crc kubenswrapper[4758]: I1203 17:31:44.632169 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerStarted","Data":"776e93219a1be154268333a75a2f1b5b16158657b88c966bea72c0bbcf594fe9"} Dec 03 17:31:44 crc kubenswrapper[4758]: I1203 17:31:44.671245 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-55zrw" podStartSLOduration=3.25131063 podStartE2EDuration="5.671205658s" podCreationTimestamp="2025-12-03 17:31:39 +0000 UTC" firstStartedPulling="2025-12-03 17:31:41.604513893 +0000 UTC m=+2156.805890804" lastFinishedPulling="2025-12-03 17:31:44.024408931 +0000 UTC m=+2159.225785832" observedRunningTime="2025-12-03 17:31:44.661277983 +0000 UTC m=+2159.862654844" watchObservedRunningTime="2025-12-03 17:31:44.671205658 +0000 UTC m=+2159.872582559" Dec 03 17:31:49 crc kubenswrapper[4758]: I1203 17:31:49.971431 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:49 crc kubenswrapper[4758]: I1203 17:31:49.972280 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:50 crc kubenswrapper[4758]: I1203 17:31:50.030122 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:50 crc kubenswrapper[4758]: I1203 17:31:50.729663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:50 crc kubenswrapper[4758]: I1203 17:31:50.789105 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-55zrw"] Dec 03 17:31:52 crc kubenswrapper[4758]: I1203 17:31:52.708142 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-55zrw" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="registry-server" containerID="cri-o://776e93219a1be154268333a75a2f1b5b16158657b88c966bea72c0bbcf594fe9" gracePeriod=2 Dec 03 17:31:55 crc kubenswrapper[4758]: I1203 17:31:55.739097 4758 generic.go:334] "Generic (PLEG): container finished" podID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerID="776e93219a1be154268333a75a2f1b5b16158657b88c966bea72c0bbcf594fe9" exitCode=0 Dec 03 17:31:55 crc kubenswrapper[4758]: I1203 17:31:55.739155 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerDied","Data":"776e93219a1be154268333a75a2f1b5b16158657b88c966bea72c0bbcf594fe9"} Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.858404 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.890975 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-catalog-content\") pod \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.891076 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sswh\" (UniqueName: \"kubernetes.io/projected/f108482f-1e58-4c76-a5ad-4ec9628b10e2-kube-api-access-5sswh\") pod \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.891140 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-utilities\") pod \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\" (UID: \"f108482f-1e58-4c76-a5ad-4ec9628b10e2\") " Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.892502 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-utilities" (OuterVolumeSpecName: "utilities") pod "f108482f-1e58-4c76-a5ad-4ec9628b10e2" (UID: "f108482f-1e58-4c76-a5ad-4ec9628b10e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.906374 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f108482f-1e58-4c76-a5ad-4ec9628b10e2-kube-api-access-5sswh" (OuterVolumeSpecName: "kube-api-access-5sswh") pod "f108482f-1e58-4c76-a5ad-4ec9628b10e2" (UID: "f108482f-1e58-4c76-a5ad-4ec9628b10e2"). InnerVolumeSpecName "kube-api-access-5sswh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.993008 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sswh\" (UniqueName: \"kubernetes.io/projected/f108482f-1e58-4c76-a5ad-4ec9628b10e2-kube-api-access-5sswh\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:56 crc kubenswrapper[4758]: I1203 17:31:56.993046 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.018947 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f108482f-1e58-4c76-a5ad-4ec9628b10e2" (UID: "f108482f-1e58-4c76-a5ad-4ec9628b10e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.094432 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f108482f-1e58-4c76-a5ad-4ec9628b10e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.774566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-55zrw" event={"ID":"f108482f-1e58-4c76-a5ad-4ec9628b10e2","Type":"ContainerDied","Data":"7bed3216b320d456171af7f23e5237d039e9c21c63d692b56ca1b099bf8d6527"} Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.774704 4758 scope.go:117] "RemoveContainer" containerID="776e93219a1be154268333a75a2f1b5b16158657b88c966bea72c0bbcf594fe9" Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.775012 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-55zrw" Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.830871 4758 scope.go:117] "RemoveContainer" containerID="1c9e9ea0db629f50b912737a32dfacb81a4dd48835fd3cc427805539374d0537" Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.847047 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-55zrw"] Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.856724 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-55zrw"] Dec 03 17:31:57 crc kubenswrapper[4758]: I1203 17:31:57.859254 4758 scope.go:117] "RemoveContainer" containerID="7c35186bcc8eca4fb9ecf88b778cbcbc5904831341eeb89603f8440b7290834f" Dec 03 17:31:59 crc kubenswrapper[4758]: I1203 17:31:59.127015 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" path="/var/lib/kubelet/pods/f108482f-1e58-4c76-a5ad-4ec9628b10e2/volumes" Dec 03 17:33:11 crc kubenswrapper[4758]: I1203 17:33:11.395568 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:33:11 crc kubenswrapper[4758]: I1203 17:33:11.396499 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:33:41 crc kubenswrapper[4758]: I1203 17:33:41.394534 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:33:41 crc kubenswrapper[4758]: I1203 17:33:41.395507 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:34:11 crc kubenswrapper[4758]: I1203 17:34:11.395027 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:34:11 crc kubenswrapper[4758]: I1203 17:34:11.395863 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:34:11 crc kubenswrapper[4758]: I1203 17:34:11.395920 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:34:11 crc kubenswrapper[4758]: I1203 17:34:11.396821 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:34:11 crc kubenswrapper[4758]: I1203 17:34:11.396883 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" gracePeriod=600 Dec 03 17:34:11 crc kubenswrapper[4758]: E1203 17:34:11.524994 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:34:12 crc kubenswrapper[4758]: I1203 17:34:12.087066 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" exitCode=0 Dec 03 17:34:12 crc kubenswrapper[4758]: I1203 17:34:12.087168 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391"} Dec 03 17:34:12 crc kubenswrapper[4758]: I1203 17:34:12.087271 4758 scope.go:117] "RemoveContainer" containerID="7bb067f35d75463b1a551aebb285d386625b1b5d3cad4de5ba7441f9f3fdcfc4" Dec 03 17:34:12 crc kubenswrapper[4758]: I1203 17:34:12.088261 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:34:12 crc kubenswrapper[4758]: E1203 17:34:12.088990 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:34:24 crc kubenswrapper[4758]: I1203 17:34:24.115209 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:34:24 crc kubenswrapper[4758]: E1203 17:34:24.116389 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:34:35 crc kubenswrapper[4758]: I1203 17:34:35.123071 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:34:35 crc kubenswrapper[4758]: E1203 17:34:35.124326 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:34:47 crc kubenswrapper[4758]: I1203 17:34:47.114884 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:34:47 crc kubenswrapper[4758]: E1203 17:34:47.116556 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:35:01 crc kubenswrapper[4758]: I1203 17:35:01.115794 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:35:01 crc kubenswrapper[4758]: E1203 17:35:01.118745 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:35:15 crc kubenswrapper[4758]: I1203 17:35:15.118489 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:35:15 crc kubenswrapper[4758]: E1203 17:35:15.119255 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:35:27 crc kubenswrapper[4758]: I1203 17:35:27.115597 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:35:27 crc kubenswrapper[4758]: E1203 17:35:27.117026 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:35:42 crc kubenswrapper[4758]: I1203 17:35:42.114505 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:35:42 crc kubenswrapper[4758]: E1203 17:35:42.115732 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:35:54 crc kubenswrapper[4758]: I1203 17:35:54.115401 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:35:54 crc kubenswrapper[4758]: E1203 17:35:54.116765 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:36:08 crc kubenswrapper[4758]: I1203 17:36:08.115042 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:36:08 crc kubenswrapper[4758]: E1203 17:36:08.115966 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:36:20 crc kubenswrapper[4758]: I1203 17:36:20.114038 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:36:20 crc kubenswrapper[4758]: E1203 17:36:20.114821 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:36:34 crc kubenswrapper[4758]: I1203 17:36:34.114727 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:36:34 crc kubenswrapper[4758]: E1203 17:36:34.115628 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:36:46 crc kubenswrapper[4758]: I1203 17:36:46.114379 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:36:46 crc kubenswrapper[4758]: E1203 17:36:46.115205 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:37:00 crc kubenswrapper[4758]: I1203 17:37:00.115182 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:37:00 crc kubenswrapper[4758]: E1203 17:37:00.116596 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:37:13 crc kubenswrapper[4758]: I1203 17:37:13.114638 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:37:13 crc kubenswrapper[4758]: E1203 17:37:13.116390 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:37:26 crc kubenswrapper[4758]: I1203 17:37:26.114827 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:37:26 crc kubenswrapper[4758]: E1203 17:37:26.115731 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:37:40 crc kubenswrapper[4758]: I1203 17:37:40.114536 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:37:40 crc kubenswrapper[4758]: E1203 17:37:40.115519 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:37:55 crc kubenswrapper[4758]: I1203 17:37:55.118575 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:37:55 crc kubenswrapper[4758]: E1203 17:37:55.119430 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:38:07 crc kubenswrapper[4758]: I1203 17:38:07.114546 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:38:07 crc kubenswrapper[4758]: E1203 17:38:07.115346 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:38:18 crc kubenswrapper[4758]: I1203 17:38:18.114314 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:38:18 crc kubenswrapper[4758]: E1203 17:38:18.115423 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:38:32 crc kubenswrapper[4758]: I1203 17:38:32.114136 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:38:32 crc kubenswrapper[4758]: E1203 17:38:32.115040 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:38:46 crc kubenswrapper[4758]: I1203 17:38:46.115210 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:38:46 crc kubenswrapper[4758]: E1203 17:38:46.116178 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:39:01 crc kubenswrapper[4758]: I1203 17:39:01.115587 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:39:01 crc kubenswrapper[4758]: E1203 17:39:01.117379 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:39:14 crc kubenswrapper[4758]: I1203 17:39:14.114160 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:39:15 crc kubenswrapper[4758]: I1203 17:39:15.261876 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"96b84012054dfd39d0601719aa99a137b8d67a7fc1d45424c1c4b75dd1e08f62"} Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.610997 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-24wbq"] Dec 03 17:39:56 crc kubenswrapper[4758]: E1203 17:39:56.613358 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="registry-server" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.613377 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="registry-server" Dec 03 17:39:56 crc kubenswrapper[4758]: E1203 17:39:56.613394 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="extract-content" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.613401 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="extract-content" Dec 03 17:39:56 crc kubenswrapper[4758]: E1203 17:39:56.613412 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="extract-utilities" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.613420 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="extract-utilities" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.613588 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f108482f-1e58-4c76-a5ad-4ec9628b10e2" containerName="registry-server" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.614952 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24wbq"] Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.615101 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.653168 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-catalog-content\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.653239 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crdqf\" (UniqueName: \"kubernetes.io/projected/2aef530d-ed23-47c6-8e05-11cec9d854ca-kube-api-access-crdqf\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.653296 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-utilities\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.754828 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-utilities\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.754898 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-catalog-content\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.754977 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crdqf\" (UniqueName: \"kubernetes.io/projected/2aef530d-ed23-47c6-8e05-11cec9d854ca-kube-api-access-crdqf\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.755494 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-utilities\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.755550 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-catalog-content\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.774820 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crdqf\" (UniqueName: \"kubernetes.io/projected/2aef530d-ed23-47c6-8e05-11cec9d854ca-kube-api-access-crdqf\") pod \"community-operators-24wbq\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:56 crc kubenswrapper[4758]: I1203 17:39:56.937893 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:39:57 crc kubenswrapper[4758]: I1203 17:39:57.455632 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24wbq"] Dec 03 17:39:57 crc kubenswrapper[4758]: I1203 17:39:57.628345 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerStarted","Data":"b7ccb8abafa02ae4efec6936aff039ab102db5d9e1e936169c6272fd42dda1a0"} Dec 03 17:39:57 crc kubenswrapper[4758]: I1203 17:39:57.628391 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerStarted","Data":"dfca9ce01e26b61ada894821dbbef303f3c3e5f7506f9e9f28901516e97b4dc6"} Dec 03 17:39:58 crc kubenswrapper[4758]: I1203 17:39:58.643183 4758 generic.go:334] "Generic (PLEG): container finished" podID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerID="b7ccb8abafa02ae4efec6936aff039ab102db5d9e1e936169c6272fd42dda1a0" exitCode=0 Dec 03 17:39:58 crc kubenswrapper[4758]: I1203 17:39:58.643373 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerDied","Data":"b7ccb8abafa02ae4efec6936aff039ab102db5d9e1e936169c6272fd42dda1a0"} Dec 03 17:39:58 crc kubenswrapper[4758]: I1203 17:39:58.647000 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:39:59 crc kubenswrapper[4758]: I1203 17:39:59.656632 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerStarted","Data":"c95c61dcb687565462b01ca9b1a1a49d265e2413191e2da157a28a09480b4eeb"} Dec 03 17:40:00 crc kubenswrapper[4758]: I1203 17:40:00.671937 4758 generic.go:334] "Generic (PLEG): container finished" podID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerID="c95c61dcb687565462b01ca9b1a1a49d265e2413191e2da157a28a09480b4eeb" exitCode=0 Dec 03 17:40:00 crc kubenswrapper[4758]: I1203 17:40:00.672063 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerDied","Data":"c95c61dcb687565462b01ca9b1a1a49d265e2413191e2da157a28a09480b4eeb"} Dec 03 17:40:01 crc kubenswrapper[4758]: I1203 17:40:01.696192 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerStarted","Data":"b0b2df34b39610ddb8b786d87ce847bab55fd36c39615f50430cbdbedbbf0c11"} Dec 03 17:40:01 crc kubenswrapper[4758]: I1203 17:40:01.723564 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-24wbq" podStartSLOduration=3.075808259 podStartE2EDuration="5.72353883s" podCreationTimestamp="2025-12-03 17:39:56 +0000 UTC" firstStartedPulling="2025-12-03 17:39:58.645510346 +0000 UTC m=+2653.846887237" lastFinishedPulling="2025-12-03 17:40:01.293240947 +0000 UTC m=+2656.494617808" observedRunningTime="2025-12-03 17:40:01.716025172 +0000 UTC m=+2656.917402073" watchObservedRunningTime="2025-12-03 17:40:01.72353883 +0000 UTC m=+2656.924915731" Dec 03 17:40:06 crc kubenswrapper[4758]: I1203 17:40:06.937951 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:40:06 crc kubenswrapper[4758]: I1203 17:40:06.940251 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:40:07 crc kubenswrapper[4758]: I1203 17:40:07.011525 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:40:07 crc kubenswrapper[4758]: I1203 17:40:07.810880 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:40:07 crc kubenswrapper[4758]: I1203 17:40:07.879159 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24wbq"] Dec 03 17:40:09 crc kubenswrapper[4758]: I1203 17:40:09.779983 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-24wbq" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="registry-server" containerID="cri-o://b0b2df34b39610ddb8b786d87ce847bab55fd36c39615f50430cbdbedbbf0c11" gracePeriod=2 Dec 03 17:40:10 crc kubenswrapper[4758]: I1203 17:40:10.796322 4758 generic.go:334] "Generic (PLEG): container finished" podID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerID="b0b2df34b39610ddb8b786d87ce847bab55fd36c39615f50430cbdbedbbf0c11" exitCode=0 Dec 03 17:40:10 crc kubenswrapper[4758]: I1203 17:40:10.796401 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerDied","Data":"b0b2df34b39610ddb8b786d87ce847bab55fd36c39615f50430cbdbedbbf0c11"} Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.349796 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.412239 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-utilities\") pod \"2aef530d-ed23-47c6-8e05-11cec9d854ca\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.412345 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crdqf\" (UniqueName: \"kubernetes.io/projected/2aef530d-ed23-47c6-8e05-11cec9d854ca-kube-api-access-crdqf\") pod \"2aef530d-ed23-47c6-8e05-11cec9d854ca\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.412394 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-catalog-content\") pod \"2aef530d-ed23-47c6-8e05-11cec9d854ca\" (UID: \"2aef530d-ed23-47c6-8e05-11cec9d854ca\") " Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.413255 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-utilities" (OuterVolumeSpecName: "utilities") pod "2aef530d-ed23-47c6-8e05-11cec9d854ca" (UID: "2aef530d-ed23-47c6-8e05-11cec9d854ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.418883 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aef530d-ed23-47c6-8e05-11cec9d854ca-kube-api-access-crdqf" (OuterVolumeSpecName: "kube-api-access-crdqf") pod "2aef530d-ed23-47c6-8e05-11cec9d854ca" (UID: "2aef530d-ed23-47c6-8e05-11cec9d854ca"). InnerVolumeSpecName "kube-api-access-crdqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.461166 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2aef530d-ed23-47c6-8e05-11cec9d854ca" (UID: "2aef530d-ed23-47c6-8e05-11cec9d854ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.513481 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crdqf\" (UniqueName: \"kubernetes.io/projected/2aef530d-ed23-47c6-8e05-11cec9d854ca-kube-api-access-crdqf\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.513517 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.513527 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aef530d-ed23-47c6-8e05-11cec9d854ca-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.814136 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24wbq" event={"ID":"2aef530d-ed23-47c6-8e05-11cec9d854ca","Type":"ContainerDied","Data":"dfca9ce01e26b61ada894821dbbef303f3c3e5f7506f9e9f28901516e97b4dc6"} Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.814243 4758 scope.go:117] "RemoveContainer" containerID="b0b2df34b39610ddb8b786d87ce847bab55fd36c39615f50430cbdbedbbf0c11" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.814310 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24wbq" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.848986 4758 scope.go:117] "RemoveContainer" containerID="c95c61dcb687565462b01ca9b1a1a49d265e2413191e2da157a28a09480b4eeb" Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.879168 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24wbq"] Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.888872 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-24wbq"] Dec 03 17:40:11 crc kubenswrapper[4758]: I1203 17:40:11.896404 4758 scope.go:117] "RemoveContainer" containerID="b7ccb8abafa02ae4efec6936aff039ab102db5d9e1e936169c6272fd42dda1a0" Dec 03 17:40:13 crc kubenswrapper[4758]: I1203 17:40:13.125449 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" path="/var/lib/kubelet/pods/2aef530d-ed23-47c6-8e05-11cec9d854ca/volumes" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.464263 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qkjzh"] Dec 03 17:40:19 crc kubenswrapper[4758]: E1203 17:40:19.465011 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="extract-content" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.465029 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="extract-content" Dec 03 17:40:19 crc kubenswrapper[4758]: E1203 17:40:19.465052 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="extract-utilities" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.465060 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="extract-utilities" Dec 03 17:40:19 crc kubenswrapper[4758]: E1203 17:40:19.465070 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="registry-server" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.465079 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="registry-server" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.465262 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aef530d-ed23-47c6-8e05-11cec9d854ca" containerName="registry-server" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.466588 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.479251 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qkjzh"] Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.535363 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-utilities\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.535435 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-catalog-content\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.535477 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqq9n\" (UniqueName: \"kubernetes.io/projected/91271f88-0d01-46e6-bc60-aadb337547f9-kube-api-access-jqq9n\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.636462 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-utilities\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.636524 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-catalog-content\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.636550 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqq9n\" (UniqueName: \"kubernetes.io/projected/91271f88-0d01-46e6-bc60-aadb337547f9-kube-api-access-jqq9n\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.637037 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-utilities\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.637112 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-catalog-content\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.663006 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqq9n\" (UniqueName: \"kubernetes.io/projected/91271f88-0d01-46e6-bc60-aadb337547f9-kube-api-access-jqq9n\") pod \"certified-operators-qkjzh\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:19 crc kubenswrapper[4758]: I1203 17:40:19.795737 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:20 crc kubenswrapper[4758]: I1203 17:40:20.271765 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qkjzh"] Dec 03 17:40:20 crc kubenswrapper[4758]: I1203 17:40:20.905048 4758 generic.go:334] "Generic (PLEG): container finished" podID="91271f88-0d01-46e6-bc60-aadb337547f9" containerID="42569b6fa5e4139b08f1393254ac51bd9f8416b5f504b1db61ae7e6f591f4b15" exitCode=0 Dec 03 17:40:20 crc kubenswrapper[4758]: I1203 17:40:20.905138 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkjzh" event={"ID":"91271f88-0d01-46e6-bc60-aadb337547f9","Type":"ContainerDied","Data":"42569b6fa5e4139b08f1393254ac51bd9f8416b5f504b1db61ae7e6f591f4b15"} Dec 03 17:40:20 crc kubenswrapper[4758]: I1203 17:40:20.907285 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkjzh" event={"ID":"91271f88-0d01-46e6-bc60-aadb337547f9","Type":"ContainerStarted","Data":"2ca57226fc8ee5e8f0118e05094141566e30eeb9551e8bc6e3268fe323fe0fe8"} Dec 03 17:40:21 crc kubenswrapper[4758]: I1203 17:40:21.923856 4758 generic.go:334] "Generic (PLEG): container finished" podID="91271f88-0d01-46e6-bc60-aadb337547f9" containerID="bb0c510112b2b8e5c9a7ce00fd6b51ea45b9fa864b07225c6a2d74b0458d33c7" exitCode=0 Dec 03 17:40:21 crc kubenswrapper[4758]: I1203 17:40:21.924386 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkjzh" event={"ID":"91271f88-0d01-46e6-bc60-aadb337547f9","Type":"ContainerDied","Data":"bb0c510112b2b8e5c9a7ce00fd6b51ea45b9fa864b07225c6a2d74b0458d33c7"} Dec 03 17:40:22 crc kubenswrapper[4758]: I1203 17:40:22.934095 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkjzh" event={"ID":"91271f88-0d01-46e6-bc60-aadb337547f9","Type":"ContainerStarted","Data":"fcd83b7314f3d55aa90a5187e30558abde7a0dda6ee93060c58d0605f02eecba"} Dec 03 17:40:22 crc kubenswrapper[4758]: I1203 17:40:22.960113 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qkjzh" podStartSLOduration=2.549473721 podStartE2EDuration="3.960095814s" podCreationTimestamp="2025-12-03 17:40:19 +0000 UTC" firstStartedPulling="2025-12-03 17:40:20.907641976 +0000 UTC m=+2676.109018877" lastFinishedPulling="2025-12-03 17:40:22.318264089 +0000 UTC m=+2677.519640970" observedRunningTime="2025-12-03 17:40:22.953639733 +0000 UTC m=+2678.155016614" watchObservedRunningTime="2025-12-03 17:40:22.960095814 +0000 UTC m=+2678.161472675" Dec 03 17:40:29 crc kubenswrapper[4758]: I1203 17:40:29.796893 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:29 crc kubenswrapper[4758]: I1203 17:40:29.797387 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:29 crc kubenswrapper[4758]: I1203 17:40:29.839352 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:30 crc kubenswrapper[4758]: I1203 17:40:30.108391 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:30 crc kubenswrapper[4758]: I1203 17:40:30.156393 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qkjzh"] Dec 03 17:40:32 crc kubenswrapper[4758]: I1203 17:40:32.006914 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qkjzh" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="registry-server" containerID="cri-o://fcd83b7314f3d55aa90a5187e30558abde7a0dda6ee93060c58d0605f02eecba" gracePeriod=2 Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.015656 4758 generic.go:334] "Generic (PLEG): container finished" podID="91271f88-0d01-46e6-bc60-aadb337547f9" containerID="fcd83b7314f3d55aa90a5187e30558abde7a0dda6ee93060c58d0605f02eecba" exitCode=0 Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.015711 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkjzh" event={"ID":"91271f88-0d01-46e6-bc60-aadb337547f9","Type":"ContainerDied","Data":"fcd83b7314f3d55aa90a5187e30558abde7a0dda6ee93060c58d0605f02eecba"} Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.111591 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.244988 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqq9n\" (UniqueName: \"kubernetes.io/projected/91271f88-0d01-46e6-bc60-aadb337547f9-kube-api-access-jqq9n\") pod \"91271f88-0d01-46e6-bc60-aadb337547f9\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.245058 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-catalog-content\") pod \"91271f88-0d01-46e6-bc60-aadb337547f9\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.245179 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-utilities\") pod \"91271f88-0d01-46e6-bc60-aadb337547f9\" (UID: \"91271f88-0d01-46e6-bc60-aadb337547f9\") " Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.246285 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-utilities" (OuterVolumeSpecName: "utilities") pod "91271f88-0d01-46e6-bc60-aadb337547f9" (UID: "91271f88-0d01-46e6-bc60-aadb337547f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.251210 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91271f88-0d01-46e6-bc60-aadb337547f9-kube-api-access-jqq9n" (OuterVolumeSpecName: "kube-api-access-jqq9n") pod "91271f88-0d01-46e6-bc60-aadb337547f9" (UID: "91271f88-0d01-46e6-bc60-aadb337547f9"). InnerVolumeSpecName "kube-api-access-jqq9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.291342 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91271f88-0d01-46e6-bc60-aadb337547f9" (UID: "91271f88-0d01-46e6-bc60-aadb337547f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.347817 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqq9n\" (UniqueName: \"kubernetes.io/projected/91271f88-0d01-46e6-bc60-aadb337547f9-kube-api-access-jqq9n\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.347857 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:33 crc kubenswrapper[4758]: I1203 17:40:33.347866 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91271f88-0d01-46e6-bc60-aadb337547f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.024626 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qkjzh" event={"ID":"91271f88-0d01-46e6-bc60-aadb337547f9","Type":"ContainerDied","Data":"2ca57226fc8ee5e8f0118e05094141566e30eeb9551e8bc6e3268fe323fe0fe8"} Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.024705 4758 scope.go:117] "RemoveContainer" containerID="fcd83b7314f3d55aa90a5187e30558abde7a0dda6ee93060c58d0605f02eecba" Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.024714 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qkjzh" Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.045732 4758 scope.go:117] "RemoveContainer" containerID="bb0c510112b2b8e5c9a7ce00fd6b51ea45b9fa864b07225c6a2d74b0458d33c7" Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.085901 4758 scope.go:117] "RemoveContainer" containerID="42569b6fa5e4139b08f1393254ac51bd9f8416b5f504b1db61ae7e6f591f4b15" Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.106171 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qkjzh"] Dec 03 17:40:34 crc kubenswrapper[4758]: I1203 17:40:34.121760 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qkjzh"] Dec 03 17:40:35 crc kubenswrapper[4758]: I1203 17:40:35.125281 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" path="/var/lib/kubelet/pods/91271f88-0d01-46e6-bc60-aadb337547f9/volumes" Dec 03 17:41:41 crc kubenswrapper[4758]: I1203 17:41:41.394344 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:41:41 crc kubenswrapper[4758]: I1203 17:41:41.394884 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:42:11 crc kubenswrapper[4758]: I1203 17:42:11.394629 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:42:11 crc kubenswrapper[4758]: I1203 17:42:11.395674 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.602860 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k8lf6"] Dec 03 17:42:33 crc kubenswrapper[4758]: E1203 17:42:33.603750 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="extract-content" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.603779 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="extract-content" Dec 03 17:42:33 crc kubenswrapper[4758]: E1203 17:42:33.603794 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="registry-server" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.603800 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="registry-server" Dec 03 17:42:33 crc kubenswrapper[4758]: E1203 17:42:33.603855 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="extract-utilities" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.603862 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="extract-utilities" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.604050 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="91271f88-0d01-46e6-bc60-aadb337547f9" containerName="registry-server" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.605409 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.617931 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8lf6"] Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.710211 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x95kc\" (UniqueName: \"kubernetes.io/projected/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-kube-api-access-x95kc\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.710278 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-utilities\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.710399 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-catalog-content\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.812127 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-catalog-content\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.812213 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x95kc\" (UniqueName: \"kubernetes.io/projected/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-kube-api-access-x95kc\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.812276 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-utilities\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.812716 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-catalog-content\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.812780 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-utilities\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.834092 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x95kc\" (UniqueName: \"kubernetes.io/projected/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-kube-api-access-x95kc\") pod \"redhat-operators-k8lf6\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:33 crc kubenswrapper[4758]: I1203 17:42:33.922101 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:42:34 crc kubenswrapper[4758]: I1203 17:42:34.347879 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8lf6"] Dec 03 17:42:34 crc kubenswrapper[4758]: I1203 17:42:34.939753 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerStarted","Data":"9dc9597dfaa9a70b9f7b01d303e34cf4abd82be636a98d6d0fae865176fd3daf"} Dec 03 17:42:35 crc kubenswrapper[4758]: I1203 17:42:35.955168 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerID="4eaef9d59be92455c648236e790762252c43c4edb3f3d4a18b84c30770398b4c" exitCode=0 Dec 03 17:42:35 crc kubenswrapper[4758]: I1203 17:42:35.955220 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerDied","Data":"4eaef9d59be92455c648236e790762252c43c4edb3f3d4a18b84c30770398b4c"} Dec 03 17:42:37 crc kubenswrapper[4758]: I1203 17:42:37.969589 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerStarted","Data":"5f51617a3c41d4b0bf9363220300bbf4ba8826bc18fcafc8976d7302a43d0ab2"} Dec 03 17:42:38 crc kubenswrapper[4758]: I1203 17:42:38.979864 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerID="5f51617a3c41d4b0bf9363220300bbf4ba8826bc18fcafc8976d7302a43d0ab2" exitCode=0 Dec 03 17:42:38 crc kubenswrapper[4758]: I1203 17:42:38.980008 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerDied","Data":"5f51617a3c41d4b0bf9363220300bbf4ba8826bc18fcafc8976d7302a43d0ab2"} Dec 03 17:42:41 crc kubenswrapper[4758]: I1203 17:42:41.395009 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:42:41 crc kubenswrapper[4758]: I1203 17:42:41.395595 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:42:41 crc kubenswrapper[4758]: I1203 17:42:41.395711 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:42:41 crc kubenswrapper[4758]: I1203 17:42:41.396736 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96b84012054dfd39d0601719aa99a137b8d67a7fc1d45424c1c4b75dd1e08f62"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:42:41 crc kubenswrapper[4758]: I1203 17:42:41.396819 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://96b84012054dfd39d0601719aa99a137b8d67a7fc1d45424c1c4b75dd1e08f62" gracePeriod=600 Dec 03 17:42:43 crc kubenswrapper[4758]: I1203 17:42:43.008519 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="96b84012054dfd39d0601719aa99a137b8d67a7fc1d45424c1c4b75dd1e08f62" exitCode=0 Dec 03 17:42:43 crc kubenswrapper[4758]: I1203 17:42:43.008857 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"96b84012054dfd39d0601719aa99a137b8d67a7fc1d45424c1c4b75dd1e08f62"} Dec 03 17:42:43 crc kubenswrapper[4758]: I1203 17:42:43.008889 4758 scope.go:117] "RemoveContainer" containerID="d3b97a751da33124a7be6c0746f3575ac43d5a045f662a05bd9afb2688327391" Dec 03 17:42:48 crc kubenswrapper[4758]: I1203 17:42:48.056248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16"} Dec 03 17:42:54 crc kubenswrapper[4758]: I1203 17:42:54.107482 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerStarted","Data":"1f787a554aa34e40402f90fd328412bc43b290eb64a1961542de16f4707e3ffc"} Dec 03 17:42:54 crc kubenswrapper[4758]: I1203 17:42:54.129070 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k8lf6" podStartSLOduration=4.098449047 podStartE2EDuration="21.129052634s" podCreationTimestamp="2025-12-03 17:42:33 +0000 UTC" firstStartedPulling="2025-12-03 17:42:35.956828543 +0000 UTC m=+2811.158205414" lastFinishedPulling="2025-12-03 17:42:52.98743214 +0000 UTC m=+2828.188809001" observedRunningTime="2025-12-03 17:42:54.123032094 +0000 UTC m=+2829.324408955" watchObservedRunningTime="2025-12-03 17:42:54.129052634 +0000 UTC m=+2829.330429495" Dec 03 17:43:03 crc kubenswrapper[4758]: I1203 17:43:03.922526 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:43:03 crc kubenswrapper[4758]: I1203 17:43:03.923362 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:43:03 crc kubenswrapper[4758]: I1203 17:43:03.992413 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:43:04 crc kubenswrapper[4758]: I1203 17:43:04.233519 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:43:04 crc kubenswrapper[4758]: I1203 17:43:04.808927 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8lf6"] Dec 03 17:43:06 crc kubenswrapper[4758]: I1203 17:43:06.215734 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k8lf6" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="registry-server" containerID="cri-o://1f787a554aa34e40402f90fd328412bc43b290eb64a1961542de16f4707e3ffc" gracePeriod=2 Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.241467 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerID="1f787a554aa34e40402f90fd328412bc43b290eb64a1961542de16f4707e3ffc" exitCode=0 Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.241546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerDied","Data":"1f787a554aa34e40402f90fd328412bc43b290eb64a1961542de16f4707e3ffc"} Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.439526 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.582731 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-utilities\") pod \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.582839 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x95kc\" (UniqueName: \"kubernetes.io/projected/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-kube-api-access-x95kc\") pod \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.582893 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-catalog-content\") pod \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\" (UID: \"cc8ae684-fbcd-4d6d-893f-3d314dfa3556\") " Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.583812 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-utilities" (OuterVolumeSpecName: "utilities") pod "cc8ae684-fbcd-4d6d-893f-3d314dfa3556" (UID: "cc8ae684-fbcd-4d6d-893f-3d314dfa3556"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.588933 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-kube-api-access-x95kc" (OuterVolumeSpecName: "kube-api-access-x95kc") pod "cc8ae684-fbcd-4d6d-893f-3d314dfa3556" (UID: "cc8ae684-fbcd-4d6d-893f-3d314dfa3556"). InnerVolumeSpecName "kube-api-access-x95kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.684087 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x95kc\" (UniqueName: \"kubernetes.io/projected/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-kube-api-access-x95kc\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.684134 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.717978 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc8ae684-fbcd-4d6d-893f-3d314dfa3556" (UID: "cc8ae684-fbcd-4d6d-893f-3d314dfa3556"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:43:07 crc kubenswrapper[4758]: I1203 17:43:07.785965 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8ae684-fbcd-4d6d-893f-3d314dfa3556-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.260031 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lf6" event={"ID":"cc8ae684-fbcd-4d6d-893f-3d314dfa3556","Type":"ContainerDied","Data":"9dc9597dfaa9a70b9f7b01d303e34cf4abd82be636a98d6d0fae865176fd3daf"} Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.260084 4758 scope.go:117] "RemoveContainer" containerID="1f787a554aa34e40402f90fd328412bc43b290eb64a1961542de16f4707e3ffc" Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.260273 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lf6" Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.292564 4758 scope.go:117] "RemoveContainer" containerID="5f51617a3c41d4b0bf9363220300bbf4ba8826bc18fcafc8976d7302a43d0ab2" Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.297805 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8lf6"] Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.302724 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k8lf6"] Dec 03 17:43:08 crc kubenswrapper[4758]: I1203 17:43:08.313897 4758 scope.go:117] "RemoveContainer" containerID="4eaef9d59be92455c648236e790762252c43c4edb3f3d4a18b84c30770398b4c" Dec 03 17:43:09 crc kubenswrapper[4758]: I1203 17:43:09.122072 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" path="/var/lib/kubelet/pods/cc8ae684-fbcd-4d6d-893f-3d314dfa3556/volumes" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.279353 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4pcpj"] Dec 03 17:43:39 crc kubenswrapper[4758]: E1203 17:43:39.281889 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="extract-content" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.282270 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="extract-content" Dec 03 17:43:39 crc kubenswrapper[4758]: E1203 17:43:39.282385 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="extract-utilities" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.282464 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="extract-utilities" Dec 03 17:43:39 crc kubenswrapper[4758]: E1203 17:43:39.282535 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="registry-server" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.282609 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="registry-server" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.282875 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8ae684-fbcd-4d6d-893f-3d314dfa3556" containerName="registry-server" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.284747 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.294928 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pcpj"] Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.479914 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-utilities\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.480017 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-catalog-content\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.480042 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b99s2\" (UniqueName: \"kubernetes.io/projected/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-kube-api-access-b99s2\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.582096 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-utilities\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.582223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-catalog-content\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.582252 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b99s2\" (UniqueName: \"kubernetes.io/projected/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-kube-api-access-b99s2\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.582370 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-utilities\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.582661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-catalog-content\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.607453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b99s2\" (UniqueName: \"kubernetes.io/projected/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-kube-api-access-b99s2\") pod \"redhat-marketplace-4pcpj\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:39 crc kubenswrapper[4758]: I1203 17:43:39.612269 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:40 crc kubenswrapper[4758]: I1203 17:43:40.061365 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pcpj"] Dec 03 17:43:40 crc kubenswrapper[4758]: I1203 17:43:40.505163 4758 generic.go:334] "Generic (PLEG): container finished" podID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerID="63ac2e29e34f08dda13517e8ee7cc10fcd79b79fe0dfe0858ab1c7c1df3080f5" exitCode=0 Dec 03 17:43:40 crc kubenswrapper[4758]: I1203 17:43:40.505212 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pcpj" event={"ID":"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec","Type":"ContainerDied","Data":"63ac2e29e34f08dda13517e8ee7cc10fcd79b79fe0dfe0858ab1c7c1df3080f5"} Dec 03 17:43:40 crc kubenswrapper[4758]: I1203 17:43:40.505243 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pcpj" event={"ID":"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec","Type":"ContainerStarted","Data":"13b9099dcbd71d3d81162181d8b69a6c172ca738f61c2c3b9f7b3c5ce604cebe"} Dec 03 17:43:41 crc kubenswrapper[4758]: I1203 17:43:41.513882 4758 generic.go:334] "Generic (PLEG): container finished" podID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerID="79ad126b8ef9c72b426027814126428275ce90952690c36008f368e726d84447" exitCode=0 Dec 03 17:43:41 crc kubenswrapper[4758]: I1203 17:43:41.513954 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pcpj" event={"ID":"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec","Type":"ContainerDied","Data":"79ad126b8ef9c72b426027814126428275ce90952690c36008f368e726d84447"} Dec 03 17:43:41 crc kubenswrapper[4758]: E1203 17:43:41.518379 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d9d10b0_3bf2_4be3_b1f8_f0fb780710ec.slice/crio-conmon-79ad126b8ef9c72b426027814126428275ce90952690c36008f368e726d84447.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:43:42 crc kubenswrapper[4758]: I1203 17:43:42.523506 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pcpj" event={"ID":"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec","Type":"ContainerStarted","Data":"d7cc901808a8c0d1ca81590278946f4373b631cd9e8c309b938fce77a071967a"} Dec 03 17:43:42 crc kubenswrapper[4758]: I1203 17:43:42.552797 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4pcpj" podStartSLOduration=2.179294457 podStartE2EDuration="3.552775309s" podCreationTimestamp="2025-12-03 17:43:39 +0000 UTC" firstStartedPulling="2025-12-03 17:43:40.520171297 +0000 UTC m=+2875.721548158" lastFinishedPulling="2025-12-03 17:43:41.893652139 +0000 UTC m=+2877.095029010" observedRunningTime="2025-12-03 17:43:42.54979683 +0000 UTC m=+2877.751173691" watchObservedRunningTime="2025-12-03 17:43:42.552775309 +0000 UTC m=+2877.754152170" Dec 03 17:43:49 crc kubenswrapper[4758]: I1203 17:43:49.613500 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:49 crc kubenswrapper[4758]: I1203 17:43:49.614151 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:49 crc kubenswrapper[4758]: I1203 17:43:49.660961 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:50 crc kubenswrapper[4758]: I1203 17:43:50.634970 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:50 crc kubenswrapper[4758]: I1203 17:43:50.685087 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pcpj"] Dec 03 17:43:52 crc kubenswrapper[4758]: I1203 17:43:52.605976 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4pcpj" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="registry-server" containerID="cri-o://d7cc901808a8c0d1ca81590278946f4373b631cd9e8c309b938fce77a071967a" gracePeriod=2 Dec 03 17:43:54 crc kubenswrapper[4758]: I1203 17:43:54.623376 4758 generic.go:334] "Generic (PLEG): container finished" podID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerID="d7cc901808a8c0d1ca81590278946f4373b631cd9e8c309b938fce77a071967a" exitCode=0 Dec 03 17:43:54 crc kubenswrapper[4758]: I1203 17:43:54.623409 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pcpj" event={"ID":"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec","Type":"ContainerDied","Data":"d7cc901808a8c0d1ca81590278946f4373b631cd9e8c309b938fce77a071967a"} Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.064448 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.120811 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-catalog-content\") pod \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.120955 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b99s2\" (UniqueName: \"kubernetes.io/projected/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-kube-api-access-b99s2\") pod \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.121029 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-utilities\") pod \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\" (UID: \"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec\") " Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.123164 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-utilities" (OuterVolumeSpecName: "utilities") pod "3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" (UID: "3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.129002 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-kube-api-access-b99s2" (OuterVolumeSpecName: "kube-api-access-b99s2") pod "3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" (UID: "3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec"). InnerVolumeSpecName "kube-api-access-b99s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.142348 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" (UID: "3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.221778 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b99s2\" (UniqueName: \"kubernetes.io/projected/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-kube-api-access-b99s2\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.221827 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.221841 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.632659 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pcpj" event={"ID":"3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec","Type":"ContainerDied","Data":"13b9099dcbd71d3d81162181d8b69a6c172ca738f61c2c3b9f7b3c5ce604cebe"} Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.632727 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pcpj" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.632731 4758 scope.go:117] "RemoveContainer" containerID="d7cc901808a8c0d1ca81590278946f4373b631cd9e8c309b938fce77a071967a" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.663273 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pcpj"] Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.665496 4758 scope.go:117] "RemoveContainer" containerID="79ad126b8ef9c72b426027814126428275ce90952690c36008f368e726d84447" Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.668180 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pcpj"] Dec 03 17:43:55 crc kubenswrapper[4758]: I1203 17:43:55.684994 4758 scope.go:117] "RemoveContainer" containerID="63ac2e29e34f08dda13517e8ee7cc10fcd79b79fe0dfe0858ab1c7c1df3080f5" Dec 03 17:43:57 crc kubenswrapper[4758]: I1203 17:43:57.128759 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" path="/var/lib/kubelet/pods/3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec/volumes" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.146148 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8"] Dec 03 17:45:00 crc kubenswrapper[4758]: E1203 17:45:00.147053 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="extract-content" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.147069 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="extract-content" Dec 03 17:45:00 crc kubenswrapper[4758]: E1203 17:45:00.147089 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="extract-utilities" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.147097 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="extract-utilities" Dec 03 17:45:00 crc kubenswrapper[4758]: E1203 17:45:00.147111 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="registry-server" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.147118 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="registry-server" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.147295 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9d10b0-3bf2-4be3-b1f8-f0fb780710ec" containerName="registry-server" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.148375 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.152120 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.152255 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.158450 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8"] Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.274993 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06f265e5-e02b-4b73-880b-d1993e88a87f-config-volume\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.275344 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06f265e5-e02b-4b73-880b-d1993e88a87f-secret-volume\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.275462 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jjlz\" (UniqueName: \"kubernetes.io/projected/06f265e5-e02b-4b73-880b-d1993e88a87f-kube-api-access-8jjlz\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.376574 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06f265e5-e02b-4b73-880b-d1993e88a87f-config-volume\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.377169 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06f265e5-e02b-4b73-880b-d1993e88a87f-secret-volume\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.377265 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jjlz\" (UniqueName: \"kubernetes.io/projected/06f265e5-e02b-4b73-880b-d1993e88a87f-kube-api-access-8jjlz\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.377553 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06f265e5-e02b-4b73-880b-d1993e88a87f-config-volume\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.390953 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06f265e5-e02b-4b73-880b-d1993e88a87f-secret-volume\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.397217 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jjlz\" (UniqueName: \"kubernetes.io/projected/06f265e5-e02b-4b73-880b-d1993e88a87f-kube-api-access-8jjlz\") pod \"collect-profiles-29413065-5pnx8\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.478597 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:00 crc kubenswrapper[4758]: I1203 17:45:00.884876 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8"] Dec 03 17:45:01 crc kubenswrapper[4758]: I1203 17:45:01.142275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" event={"ID":"06f265e5-e02b-4b73-880b-d1993e88a87f","Type":"ContainerStarted","Data":"9b2289fc2ae3834503c845da9abb3a38e925e6829a3dee6f67bef5a7b8aaf298"} Dec 03 17:45:01 crc kubenswrapper[4758]: I1203 17:45:01.142317 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" event={"ID":"06f265e5-e02b-4b73-880b-d1993e88a87f","Type":"ContainerStarted","Data":"7a3c7f2403a677264385ee0c420907bc52bf58179a3337e5fd44ec4cede06af2"} Dec 03 17:45:01 crc kubenswrapper[4758]: I1203 17:45:01.159180 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" podStartSLOduration=1.159152481 podStartE2EDuration="1.159152481s" podCreationTimestamp="2025-12-03 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:45:01.154716754 +0000 UTC m=+2956.356093645" watchObservedRunningTime="2025-12-03 17:45:01.159152481 +0000 UTC m=+2956.360529342" Dec 03 17:45:02 crc kubenswrapper[4758]: I1203 17:45:02.150562 4758 generic.go:334] "Generic (PLEG): container finished" podID="06f265e5-e02b-4b73-880b-d1993e88a87f" containerID="9b2289fc2ae3834503c845da9abb3a38e925e6829a3dee6f67bef5a7b8aaf298" exitCode=0 Dec 03 17:45:02 crc kubenswrapper[4758]: I1203 17:45:02.150607 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" event={"ID":"06f265e5-e02b-4b73-880b-d1993e88a87f","Type":"ContainerDied","Data":"9b2289fc2ae3834503c845da9abb3a38e925e6829a3dee6f67bef5a7b8aaf298"} Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.424635 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.521438 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06f265e5-e02b-4b73-880b-d1993e88a87f-config-volume\") pod \"06f265e5-e02b-4b73-880b-d1993e88a87f\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.521723 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06f265e5-e02b-4b73-880b-d1993e88a87f-secret-volume\") pod \"06f265e5-e02b-4b73-880b-d1993e88a87f\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.521784 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jjlz\" (UniqueName: \"kubernetes.io/projected/06f265e5-e02b-4b73-880b-d1993e88a87f-kube-api-access-8jjlz\") pod \"06f265e5-e02b-4b73-880b-d1993e88a87f\" (UID: \"06f265e5-e02b-4b73-880b-d1993e88a87f\") " Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.521970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f265e5-e02b-4b73-880b-d1993e88a87f-config-volume" (OuterVolumeSpecName: "config-volume") pod "06f265e5-e02b-4b73-880b-d1993e88a87f" (UID: "06f265e5-e02b-4b73-880b-d1993e88a87f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.526619 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f265e5-e02b-4b73-880b-d1993e88a87f-kube-api-access-8jjlz" (OuterVolumeSpecName: "kube-api-access-8jjlz") pod "06f265e5-e02b-4b73-880b-d1993e88a87f" (UID: "06f265e5-e02b-4b73-880b-d1993e88a87f"). InnerVolumeSpecName "kube-api-access-8jjlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.528805 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f265e5-e02b-4b73-880b-d1993e88a87f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "06f265e5-e02b-4b73-880b-d1993e88a87f" (UID: "06f265e5-e02b-4b73-880b-d1993e88a87f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.623732 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06f265e5-e02b-4b73-880b-d1993e88a87f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.623766 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06f265e5-e02b-4b73-880b-d1993e88a87f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:03 crc kubenswrapper[4758]: I1203 17:45:03.623776 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jjlz\" (UniqueName: \"kubernetes.io/projected/06f265e5-e02b-4b73-880b-d1993e88a87f-kube-api-access-8jjlz\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:04 crc kubenswrapper[4758]: I1203 17:45:04.165170 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" event={"ID":"06f265e5-e02b-4b73-880b-d1993e88a87f","Type":"ContainerDied","Data":"7a3c7f2403a677264385ee0c420907bc52bf58179a3337e5fd44ec4cede06af2"} Dec 03 17:45:04 crc kubenswrapper[4758]: I1203 17:45:04.165211 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a3c7f2403a677264385ee0c420907bc52bf58179a3337e5fd44ec4cede06af2" Dec 03 17:45:04 crc kubenswrapper[4758]: I1203 17:45:04.165223 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8" Dec 03 17:45:04 crc kubenswrapper[4758]: I1203 17:45:04.228541 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf"] Dec 03 17:45:04 crc kubenswrapper[4758]: I1203 17:45:04.233739 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-ck5lf"] Dec 03 17:45:05 crc kubenswrapper[4758]: I1203 17:45:05.122187 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a25891-06ed-4b49-ab23-6551b6903199" path="/var/lib/kubelet/pods/58a25891-06ed-4b49-ab23-6551b6903199/volumes" Dec 03 17:45:09 crc kubenswrapper[4758]: I1203 17:45:09.791791 4758 scope.go:117] "RemoveContainer" containerID="53a2bbd9d7f93ea90625056446dfd2ea459065e55c56c4f570c4d2ce7cf5ee74" Dec 03 17:45:11 crc kubenswrapper[4758]: I1203 17:45:11.394412 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:45:11 crc kubenswrapper[4758]: I1203 17:45:11.394993 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:45:41 crc kubenswrapper[4758]: I1203 17:45:41.394716 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:45:41 crc kubenswrapper[4758]: I1203 17:45:41.395368 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.394707 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.395302 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.395353 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.396090 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.396164 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" gracePeriod=600 Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.716161 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" exitCode=0 Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.716222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16"} Dec 03 17:46:11 crc kubenswrapper[4758]: I1203 17:46:11.716276 4758 scope.go:117] "RemoveContainer" containerID="96b84012054dfd39d0601719aa99a137b8d67a7fc1d45424c1c4b75dd1e08f62" Dec 03 17:46:12 crc kubenswrapper[4758]: E1203 17:46:12.199902 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:46:12 crc kubenswrapper[4758]: I1203 17:46:12.726030 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:46:12 crc kubenswrapper[4758]: E1203 17:46:12.726283 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:46:25 crc kubenswrapper[4758]: I1203 17:46:25.117694 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:46:25 crc kubenswrapper[4758]: E1203 17:46:25.118808 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:46:40 crc kubenswrapper[4758]: I1203 17:46:40.114824 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:46:40 crc kubenswrapper[4758]: E1203 17:46:40.115667 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:46:55 crc kubenswrapper[4758]: I1203 17:46:55.115027 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:46:55 crc kubenswrapper[4758]: E1203 17:46:55.115981 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:47:06 crc kubenswrapper[4758]: I1203 17:47:06.115611 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:47:06 crc kubenswrapper[4758]: E1203 17:47:06.117184 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:47:19 crc kubenswrapper[4758]: I1203 17:47:19.115445 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:47:19 crc kubenswrapper[4758]: E1203 17:47:19.116291 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:47:30 crc kubenswrapper[4758]: I1203 17:47:30.114177 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:47:30 crc kubenswrapper[4758]: E1203 17:47:30.115177 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:47:43 crc kubenswrapper[4758]: I1203 17:47:43.115277 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:47:43 crc kubenswrapper[4758]: E1203 17:47:43.116470 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:47:57 crc kubenswrapper[4758]: I1203 17:47:57.114118 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:47:57 crc kubenswrapper[4758]: E1203 17:47:57.115182 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:48:09 crc kubenswrapper[4758]: I1203 17:48:09.114278 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:48:09 crc kubenswrapper[4758]: E1203 17:48:09.115011 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:48:24 crc kubenswrapper[4758]: I1203 17:48:24.115472 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:48:24 crc kubenswrapper[4758]: E1203 17:48:24.116273 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:48:35 crc kubenswrapper[4758]: I1203 17:48:35.118595 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:48:35 crc kubenswrapper[4758]: E1203 17:48:35.122000 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:48:48 crc kubenswrapper[4758]: I1203 17:48:48.126590 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:48:48 crc kubenswrapper[4758]: E1203 17:48:48.127560 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:49:01 crc kubenswrapper[4758]: I1203 17:49:01.113986 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:49:01 crc kubenswrapper[4758]: E1203 17:49:01.114811 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:49:12 crc kubenswrapper[4758]: I1203 17:49:12.115248 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:49:12 crc kubenswrapper[4758]: E1203 17:49:12.116922 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:49:23 crc kubenswrapper[4758]: I1203 17:49:23.114235 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:49:23 crc kubenswrapper[4758]: E1203 17:49:23.114766 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:49:37 crc kubenswrapper[4758]: I1203 17:49:37.114468 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:49:37 crc kubenswrapper[4758]: E1203 17:49:37.115513 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:49:48 crc kubenswrapper[4758]: I1203 17:49:48.114599 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:49:48 crc kubenswrapper[4758]: E1203 17:49:48.115302 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.243142 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v5cwc"] Dec 03 17:49:58 crc kubenswrapper[4758]: E1203 17:49:58.243996 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f265e5-e02b-4b73-880b-d1993e88a87f" containerName="collect-profiles" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.244016 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f265e5-e02b-4b73-880b-d1993e88a87f" containerName="collect-profiles" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.244278 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f265e5-e02b-4b73-880b-d1993e88a87f" containerName="collect-profiles" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.245365 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.261943 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5cwc"] Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.328860 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-catalog-content\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.329007 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdf5k\" (UniqueName: \"kubernetes.io/projected/e2273212-282a-4ddc-a771-216b34b6b652-kube-api-access-vdf5k\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.329047 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-utilities\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.431071 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdf5k\" (UniqueName: \"kubernetes.io/projected/e2273212-282a-4ddc-a771-216b34b6b652-kube-api-access-vdf5k\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.431121 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-utilities\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.431174 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-catalog-content\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.431730 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-catalog-content\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.432355 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-utilities\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.464654 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdf5k\" (UniqueName: \"kubernetes.io/projected/e2273212-282a-4ddc-a771-216b34b6b652-kube-api-access-vdf5k\") pod \"community-operators-v5cwc\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:58 crc kubenswrapper[4758]: I1203 17:49:58.571325 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:49:59 crc kubenswrapper[4758]: I1203 17:49:59.045011 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5cwc"] Dec 03 17:49:59 crc kubenswrapper[4758]: I1203 17:49:59.413414 4758 generic.go:334] "Generic (PLEG): container finished" podID="e2273212-282a-4ddc-a771-216b34b6b652" containerID="69ccae3ccc4b67df473ef681138055c2b5e92317c889a9f669dd558470991c15" exitCode=0 Dec 03 17:49:59 crc kubenswrapper[4758]: I1203 17:49:59.413498 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5cwc" event={"ID":"e2273212-282a-4ddc-a771-216b34b6b652","Type":"ContainerDied","Data":"69ccae3ccc4b67df473ef681138055c2b5e92317c889a9f669dd558470991c15"} Dec 03 17:49:59 crc kubenswrapper[4758]: I1203 17:49:59.413720 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5cwc" event={"ID":"e2273212-282a-4ddc-a771-216b34b6b652","Type":"ContainerStarted","Data":"1257e9e2c89970a69d26ec44764090d126d56c8e0e2af348237c75a74542954e"} Dec 03 17:49:59 crc kubenswrapper[4758]: I1203 17:49:59.415448 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:50:00 crc kubenswrapper[4758]: I1203 17:50:00.422972 4758 generic.go:334] "Generic (PLEG): container finished" podID="e2273212-282a-4ddc-a771-216b34b6b652" containerID="10ec827d058e7929f15bacb4fc96bbc5123cc5bf65676ba7a892e567b4dcd1e3" exitCode=0 Dec 03 17:50:00 crc kubenswrapper[4758]: I1203 17:50:00.423023 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5cwc" event={"ID":"e2273212-282a-4ddc-a771-216b34b6b652","Type":"ContainerDied","Data":"10ec827d058e7929f15bacb4fc96bbc5123cc5bf65676ba7a892e567b4dcd1e3"} Dec 03 17:50:02 crc kubenswrapper[4758]: I1203 17:50:02.115012 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:50:02 crc kubenswrapper[4758]: E1203 17:50:02.115525 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:50:02 crc kubenswrapper[4758]: I1203 17:50:02.439591 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5cwc" event={"ID":"e2273212-282a-4ddc-a771-216b34b6b652","Type":"ContainerStarted","Data":"6eb6d3f2b5bbf81b0d78caf12a876da3c949eeda97366a189e1e6b00ed482aac"} Dec 03 17:50:02 crc kubenswrapper[4758]: I1203 17:50:02.458052 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v5cwc" podStartSLOduration=2.148747661 podStartE2EDuration="4.458030526s" podCreationTimestamp="2025-12-03 17:49:58 +0000 UTC" firstStartedPulling="2025-12-03 17:49:59.415167031 +0000 UTC m=+3254.616543892" lastFinishedPulling="2025-12-03 17:50:01.724449896 +0000 UTC m=+3256.925826757" observedRunningTime="2025-12-03 17:50:02.456825265 +0000 UTC m=+3257.658202136" watchObservedRunningTime="2025-12-03 17:50:02.458030526 +0000 UTC m=+3257.659407387" Dec 03 17:50:08 crc kubenswrapper[4758]: I1203 17:50:08.571619 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:50:08 crc kubenswrapper[4758]: I1203 17:50:08.572235 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:50:08 crc kubenswrapper[4758]: I1203 17:50:08.614728 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:50:09 crc kubenswrapper[4758]: I1203 17:50:09.531289 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:50:09 crc kubenswrapper[4758]: I1203 17:50:09.575930 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5cwc"] Dec 03 17:50:11 crc kubenswrapper[4758]: I1203 17:50:11.505294 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v5cwc" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="registry-server" containerID="cri-o://6eb6d3f2b5bbf81b0d78caf12a876da3c949eeda97366a189e1e6b00ed482aac" gracePeriod=2 Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.514772 4758 generic.go:334] "Generic (PLEG): container finished" podID="e2273212-282a-4ddc-a771-216b34b6b652" containerID="6eb6d3f2b5bbf81b0d78caf12a876da3c949eeda97366a189e1e6b00ed482aac" exitCode=0 Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.514826 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5cwc" event={"ID":"e2273212-282a-4ddc-a771-216b34b6b652","Type":"ContainerDied","Data":"6eb6d3f2b5bbf81b0d78caf12a876da3c949eeda97366a189e1e6b00ed482aac"} Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.514893 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5cwc" event={"ID":"e2273212-282a-4ddc-a771-216b34b6b652","Type":"ContainerDied","Data":"1257e9e2c89970a69d26ec44764090d126d56c8e0e2af348237c75a74542954e"} Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.514909 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1257e9e2c89970a69d26ec44764090d126d56c8e0e2af348237c75a74542954e" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.534278 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.626574 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-utilities\") pod \"e2273212-282a-4ddc-a771-216b34b6b652\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.626770 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-catalog-content\") pod \"e2273212-282a-4ddc-a771-216b34b6b652\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.626831 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdf5k\" (UniqueName: \"kubernetes.io/projected/e2273212-282a-4ddc-a771-216b34b6b652-kube-api-access-vdf5k\") pod \"e2273212-282a-4ddc-a771-216b34b6b652\" (UID: \"e2273212-282a-4ddc-a771-216b34b6b652\") " Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.627971 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-utilities" (OuterVolumeSpecName: "utilities") pod "e2273212-282a-4ddc-a771-216b34b6b652" (UID: "e2273212-282a-4ddc-a771-216b34b6b652"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.634729 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2273212-282a-4ddc-a771-216b34b6b652-kube-api-access-vdf5k" (OuterVolumeSpecName: "kube-api-access-vdf5k") pod "e2273212-282a-4ddc-a771-216b34b6b652" (UID: "e2273212-282a-4ddc-a771-216b34b6b652"). InnerVolumeSpecName "kube-api-access-vdf5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.677377 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2273212-282a-4ddc-a771-216b34b6b652" (UID: "e2273212-282a-4ddc-a771-216b34b6b652"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.729433 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.729484 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdf5k\" (UniqueName: \"kubernetes.io/projected/e2273212-282a-4ddc-a771-216b34b6b652-kube-api-access-vdf5k\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:12 crc kubenswrapper[4758]: I1203 17:50:12.729507 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2273212-282a-4ddc-a771-216b34b6b652-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:13 crc kubenswrapper[4758]: I1203 17:50:13.520081 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5cwc" Dec 03 17:50:13 crc kubenswrapper[4758]: I1203 17:50:13.544356 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5cwc"] Dec 03 17:50:13 crc kubenswrapper[4758]: I1203 17:50:13.550296 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v5cwc"] Dec 03 17:50:14 crc kubenswrapper[4758]: I1203 17:50:14.115373 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:50:14 crc kubenswrapper[4758]: E1203 17:50:14.115741 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:50:15 crc kubenswrapper[4758]: I1203 17:50:15.123964 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2273212-282a-4ddc-a771-216b34b6b652" path="/var/lib/kubelet/pods/e2273212-282a-4ddc-a771-216b34b6b652/volumes" Dec 03 17:50:25 crc kubenswrapper[4758]: I1203 17:50:25.117781 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:50:25 crc kubenswrapper[4758]: E1203 17:50:25.118501 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:50:37 crc kubenswrapper[4758]: I1203 17:50:37.114540 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:50:37 crc kubenswrapper[4758]: E1203 17:50:37.115480 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:50:52 crc kubenswrapper[4758]: I1203 17:50:52.115063 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:50:52 crc kubenswrapper[4758]: E1203 17:50:52.116548 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.775417 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbgp8"] Dec 03 17:51:03 crc kubenswrapper[4758]: E1203 17:51:03.776417 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="extract-content" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.776433 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="extract-content" Dec 03 17:51:03 crc kubenswrapper[4758]: E1203 17:51:03.776447 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="extract-utilities" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.776454 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="extract-utilities" Dec 03 17:51:03 crc kubenswrapper[4758]: E1203 17:51:03.776484 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="registry-server" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.776492 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="registry-server" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.776754 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2273212-282a-4ddc-a771-216b34b6b652" containerName="registry-server" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.778275 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.784653 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbgp8"] Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.858506 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-catalog-content\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.858583 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-utilities\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.858608 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnpf9\" (UniqueName: \"kubernetes.io/projected/e6e164a6-aca5-4f28-b211-06300a5c791d-kube-api-access-pnpf9\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.960258 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-utilities\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.960342 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnpf9\" (UniqueName: \"kubernetes.io/projected/e6e164a6-aca5-4f28-b211-06300a5c791d-kube-api-access-pnpf9\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.960452 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-catalog-content\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.960786 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-utilities\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.961120 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-catalog-content\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:03 crc kubenswrapper[4758]: I1203 17:51:03.992881 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnpf9\" (UniqueName: \"kubernetes.io/projected/e6e164a6-aca5-4f28-b211-06300a5c791d-kube-api-access-pnpf9\") pod \"certified-operators-dbgp8\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:04 crc kubenswrapper[4758]: I1203 17:51:04.103870 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:04 crc kubenswrapper[4758]: I1203 17:51:04.582443 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbgp8"] Dec 03 17:51:04 crc kubenswrapper[4758]: I1203 17:51:04.881960 4758 generic.go:334] "Generic (PLEG): container finished" podID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerID="169fe79f418800fca79085c0b421aa5de0ec76e8492bf80e0cccd74dbb9576b3" exitCode=0 Dec 03 17:51:04 crc kubenswrapper[4758]: I1203 17:51:04.882003 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerDied","Data":"169fe79f418800fca79085c0b421aa5de0ec76e8492bf80e0cccd74dbb9576b3"} Dec 03 17:51:04 crc kubenswrapper[4758]: I1203 17:51:04.882028 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerStarted","Data":"a25419fa0564b3310840a2284bfa17897727cd38536ff14ff840ad04bb4f81fb"} Dec 03 17:51:05 crc kubenswrapper[4758]: I1203 17:51:05.120378 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:51:05 crc kubenswrapper[4758]: E1203 17:51:05.120879 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:51:05 crc kubenswrapper[4758]: I1203 17:51:05.889838 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerStarted","Data":"19b65b3338a411cceac3ea42f42efb4530629d94e5167c30ca6e889c15f1da25"} Dec 03 17:51:06 crc kubenswrapper[4758]: I1203 17:51:06.897509 4758 generic.go:334] "Generic (PLEG): container finished" podID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerID="19b65b3338a411cceac3ea42f42efb4530629d94e5167c30ca6e889c15f1da25" exitCode=0 Dec 03 17:51:06 crc kubenswrapper[4758]: I1203 17:51:06.897643 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerDied","Data":"19b65b3338a411cceac3ea42f42efb4530629d94e5167c30ca6e889c15f1da25"} Dec 03 17:51:08 crc kubenswrapper[4758]: I1203 17:51:08.931726 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerStarted","Data":"d187bfa3f071599ca84ca581768990b29f739fd85700da2b545efd7309f114e7"} Dec 03 17:51:08 crc kubenswrapper[4758]: I1203 17:51:08.955248 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbgp8" podStartSLOduration=3.555526241 podStartE2EDuration="5.95522873s" podCreationTimestamp="2025-12-03 17:51:03 +0000 UTC" firstStartedPulling="2025-12-03 17:51:04.883268389 +0000 UTC m=+3320.084645250" lastFinishedPulling="2025-12-03 17:51:07.282970878 +0000 UTC m=+3322.484347739" observedRunningTime="2025-12-03 17:51:08.948623855 +0000 UTC m=+3324.150000716" watchObservedRunningTime="2025-12-03 17:51:08.95522873 +0000 UTC m=+3324.156605591" Dec 03 17:51:14 crc kubenswrapper[4758]: I1203 17:51:14.104968 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:14 crc kubenswrapper[4758]: I1203 17:51:14.105298 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:14 crc kubenswrapper[4758]: I1203 17:51:14.153345 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:15 crc kubenswrapper[4758]: I1203 17:51:15.020580 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:15 crc kubenswrapper[4758]: I1203 17:51:15.066543 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbgp8"] Dec 03 17:51:16 crc kubenswrapper[4758]: I1203 17:51:16.994859 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbgp8" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="registry-server" containerID="cri-o://d187bfa3f071599ca84ca581768990b29f739fd85700da2b545efd7309f114e7" gracePeriod=2 Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.022387 4758 generic.go:334] "Generic (PLEG): container finished" podID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerID="d187bfa3f071599ca84ca581768990b29f739fd85700da2b545efd7309f114e7" exitCode=0 Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.022447 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerDied","Data":"d187bfa3f071599ca84ca581768990b29f739fd85700da2b545efd7309f114e7"} Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.473327 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.568776 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnpf9\" (UniqueName: \"kubernetes.io/projected/e6e164a6-aca5-4f28-b211-06300a5c791d-kube-api-access-pnpf9\") pod \"e6e164a6-aca5-4f28-b211-06300a5c791d\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.568846 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-utilities\") pod \"e6e164a6-aca5-4f28-b211-06300a5c791d\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.569005 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-catalog-content\") pod \"e6e164a6-aca5-4f28-b211-06300a5c791d\" (UID: \"e6e164a6-aca5-4f28-b211-06300a5c791d\") " Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.569743 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-utilities" (OuterVolumeSpecName: "utilities") pod "e6e164a6-aca5-4f28-b211-06300a5c791d" (UID: "e6e164a6-aca5-4f28-b211-06300a5c791d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.574486 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e164a6-aca5-4f28-b211-06300a5c791d-kube-api-access-pnpf9" (OuterVolumeSpecName: "kube-api-access-pnpf9") pod "e6e164a6-aca5-4f28-b211-06300a5c791d" (UID: "e6e164a6-aca5-4f28-b211-06300a5c791d"). InnerVolumeSpecName "kube-api-access-pnpf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.622471 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6e164a6-aca5-4f28-b211-06300a5c791d" (UID: "e6e164a6-aca5-4f28-b211-06300a5c791d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.670468 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.670504 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnpf9\" (UniqueName: \"kubernetes.io/projected/e6e164a6-aca5-4f28-b211-06300a5c791d-kube-api-access-pnpf9\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:18 crc kubenswrapper[4758]: I1203 17:51:18.670513 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e164a6-aca5-4f28-b211-06300a5c791d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.035161 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbgp8" event={"ID":"e6e164a6-aca5-4f28-b211-06300a5c791d","Type":"ContainerDied","Data":"a25419fa0564b3310840a2284bfa17897727cd38536ff14ff840ad04bb4f81fb"} Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.035253 4758 scope.go:117] "RemoveContainer" containerID="d187bfa3f071599ca84ca581768990b29f739fd85700da2b545efd7309f114e7" Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.035272 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbgp8" Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.055357 4758 scope.go:117] "RemoveContainer" containerID="19b65b3338a411cceac3ea42f42efb4530629d94e5167c30ca6e889c15f1da25" Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.079232 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbgp8"] Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.089008 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbgp8"] Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.089436 4758 scope.go:117] "RemoveContainer" containerID="169fe79f418800fca79085c0b421aa5de0ec76e8492bf80e0cccd74dbb9576b3" Dec 03 17:51:19 crc kubenswrapper[4758]: I1203 17:51:19.123348 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" path="/var/lib/kubelet/pods/e6e164a6-aca5-4f28-b211-06300a5c791d/volumes" Dec 03 17:51:20 crc kubenswrapper[4758]: I1203 17:51:20.115064 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:51:21 crc kubenswrapper[4758]: I1203 17:51:21.050791 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"2ddb3a561d726a22607fe93840c810114243f0107c35ca00e535adbcce402682"} Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.565432 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vqsgl"] Dec 03 17:52:40 crc kubenswrapper[4758]: E1203 17:52:40.566399 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="extract-content" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.566417 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="extract-content" Dec 03 17:52:40 crc kubenswrapper[4758]: E1203 17:52:40.566438 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="registry-server" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.566447 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="registry-server" Dec 03 17:52:40 crc kubenswrapper[4758]: E1203 17:52:40.566466 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="extract-utilities" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.566475 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="extract-utilities" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.566661 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e164a6-aca5-4f28-b211-06300a5c791d" containerName="registry-server" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.568047 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.580043 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vqsgl"] Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.665554 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghnq\" (UniqueName: \"kubernetes.io/projected/e515e8b1-dbd9-4207-a9ba-39f87a426014-kube-api-access-5ghnq\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.665666 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-catalog-content\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.665725 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-utilities\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.767402 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-utilities\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.767484 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghnq\" (UniqueName: \"kubernetes.io/projected/e515e8b1-dbd9-4207-a9ba-39f87a426014-kube-api-access-5ghnq\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.767557 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-catalog-content\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.768296 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-catalog-content\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.769296 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-utilities\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.790003 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghnq\" (UniqueName: \"kubernetes.io/projected/e515e8b1-dbd9-4207-a9ba-39f87a426014-kube-api-access-5ghnq\") pod \"redhat-operators-vqsgl\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:40 crc kubenswrapper[4758]: I1203 17:52:40.903550 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:41 crc kubenswrapper[4758]: I1203 17:52:41.340448 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vqsgl"] Dec 03 17:52:41 crc kubenswrapper[4758]: I1203 17:52:41.647019 4758 generic.go:334] "Generic (PLEG): container finished" podID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerID="fc3a36f28961ac65fdd5bb942be726837ae4c39a592984b9fd728dd5065beb56" exitCode=0 Dec 03 17:52:41 crc kubenswrapper[4758]: I1203 17:52:41.647062 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerDied","Data":"fc3a36f28961ac65fdd5bb942be726837ae4c39a592984b9fd728dd5065beb56"} Dec 03 17:52:41 crc kubenswrapper[4758]: I1203 17:52:41.647089 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerStarted","Data":"3586fb6cdde5668c9ae9a4e91e2b7865a4b2777b1e85e951adc9196d53a0f330"} Dec 03 17:52:42 crc kubenswrapper[4758]: I1203 17:52:42.655838 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerStarted","Data":"38b2ec38520dddbe2e12ca209c2bb1c3f5cc5c06ca71b1d6e48707ac36809e0b"} Dec 03 17:52:43 crc kubenswrapper[4758]: I1203 17:52:43.664500 4758 generic.go:334] "Generic (PLEG): container finished" podID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerID="38b2ec38520dddbe2e12ca209c2bb1c3f5cc5c06ca71b1d6e48707ac36809e0b" exitCode=0 Dec 03 17:52:43 crc kubenswrapper[4758]: I1203 17:52:43.664548 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerDied","Data":"38b2ec38520dddbe2e12ca209c2bb1c3f5cc5c06ca71b1d6e48707ac36809e0b"} Dec 03 17:52:44 crc kubenswrapper[4758]: I1203 17:52:44.676541 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerStarted","Data":"24af2d15c224a0e3c8295098d8b0a947da1fa3d0f5fa2a98167a98a7a16860cd"} Dec 03 17:52:44 crc kubenswrapper[4758]: I1203 17:52:44.697588 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vqsgl" podStartSLOduration=2.181770647 podStartE2EDuration="4.697570464s" podCreationTimestamp="2025-12-03 17:52:40 +0000 UTC" firstStartedPulling="2025-12-03 17:52:41.648560955 +0000 UTC m=+3416.849937816" lastFinishedPulling="2025-12-03 17:52:44.164360772 +0000 UTC m=+3419.365737633" observedRunningTime="2025-12-03 17:52:44.691892313 +0000 UTC m=+3419.893269194" watchObservedRunningTime="2025-12-03 17:52:44.697570464 +0000 UTC m=+3419.898947325" Dec 03 17:52:50 crc kubenswrapper[4758]: I1203 17:52:50.904182 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:50 crc kubenswrapper[4758]: I1203 17:52:50.904801 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:50 crc kubenswrapper[4758]: I1203 17:52:50.964281 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:51 crc kubenswrapper[4758]: I1203 17:52:51.774376 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:51 crc kubenswrapper[4758]: I1203 17:52:51.813902 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vqsgl"] Dec 03 17:52:53 crc kubenswrapper[4758]: I1203 17:52:53.746417 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vqsgl" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="registry-server" containerID="cri-o://24af2d15c224a0e3c8295098d8b0a947da1fa3d0f5fa2a98167a98a7a16860cd" gracePeriod=2 Dec 03 17:52:55 crc kubenswrapper[4758]: I1203 17:52:55.762357 4758 generic.go:334] "Generic (PLEG): container finished" podID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerID="24af2d15c224a0e3c8295098d8b0a947da1fa3d0f5fa2a98167a98a7a16860cd" exitCode=0 Dec 03 17:52:55 crc kubenswrapper[4758]: I1203 17:52:55.762509 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerDied","Data":"24af2d15c224a0e3c8295098d8b0a947da1fa3d0f5fa2a98167a98a7a16860cd"} Dec 03 17:52:55 crc kubenswrapper[4758]: I1203 17:52:55.964808 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.088041 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ghnq\" (UniqueName: \"kubernetes.io/projected/e515e8b1-dbd9-4207-a9ba-39f87a426014-kube-api-access-5ghnq\") pod \"e515e8b1-dbd9-4207-a9ba-39f87a426014\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.088147 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-utilities\") pod \"e515e8b1-dbd9-4207-a9ba-39f87a426014\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.088201 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-catalog-content\") pod \"e515e8b1-dbd9-4207-a9ba-39f87a426014\" (UID: \"e515e8b1-dbd9-4207-a9ba-39f87a426014\") " Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.089925 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-utilities" (OuterVolumeSpecName: "utilities") pod "e515e8b1-dbd9-4207-a9ba-39f87a426014" (UID: "e515e8b1-dbd9-4207-a9ba-39f87a426014"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.094744 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e515e8b1-dbd9-4207-a9ba-39f87a426014-kube-api-access-5ghnq" (OuterVolumeSpecName: "kube-api-access-5ghnq") pod "e515e8b1-dbd9-4207-a9ba-39f87a426014" (UID: "e515e8b1-dbd9-4207-a9ba-39f87a426014"). InnerVolumeSpecName "kube-api-access-5ghnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.190670 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ghnq\" (UniqueName: \"kubernetes.io/projected/e515e8b1-dbd9-4207-a9ba-39f87a426014-kube-api-access-5ghnq\") on node \"crc\" DevicePath \"\"" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.190755 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.211854 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e515e8b1-dbd9-4207-a9ba-39f87a426014" (UID: "e515e8b1-dbd9-4207-a9ba-39f87a426014"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.292605 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e8b1-dbd9-4207-a9ba-39f87a426014-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.771943 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqsgl" event={"ID":"e515e8b1-dbd9-4207-a9ba-39f87a426014","Type":"ContainerDied","Data":"3586fb6cdde5668c9ae9a4e91e2b7865a4b2777b1e85e951adc9196d53a0f330"} Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.772013 4758 scope.go:117] "RemoveContainer" containerID="24af2d15c224a0e3c8295098d8b0a947da1fa3d0f5fa2a98167a98a7a16860cd" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.772034 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqsgl" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.790962 4758 scope.go:117] "RemoveContainer" containerID="38b2ec38520dddbe2e12ca209c2bb1c3f5cc5c06ca71b1d6e48707ac36809e0b" Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.805755 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vqsgl"] Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.812080 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vqsgl"] Dec 03 17:52:56 crc kubenswrapper[4758]: I1203 17:52:56.831236 4758 scope.go:117] "RemoveContainer" containerID="fc3a36f28961ac65fdd5bb942be726837ae4c39a592984b9fd728dd5065beb56" Dec 03 17:52:57 crc kubenswrapper[4758]: I1203 17:52:57.140846 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" path="/var/lib/kubelet/pods/e515e8b1-dbd9-4207-a9ba-39f87a426014/volumes" Dec 03 17:53:41 crc kubenswrapper[4758]: I1203 17:53:41.394180 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:53:41 crc kubenswrapper[4758]: I1203 17:53:41.394674 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.649616 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-956sh"] Dec 03 17:53:44 crc kubenswrapper[4758]: E1203 17:53:44.650613 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="extract-content" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.650637 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="extract-content" Dec 03 17:53:44 crc kubenswrapper[4758]: E1203 17:53:44.650653 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="registry-server" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.650666 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="registry-server" Dec 03 17:53:44 crc kubenswrapper[4758]: E1203 17:53:44.650738 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="extract-utilities" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.650750 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="extract-utilities" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.650980 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e515e8b1-dbd9-4207-a9ba-39f87a426014" containerName="registry-server" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.652720 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.659409 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-956sh"] Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.774585 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-catalog-content\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.774644 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-utilities\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.774787 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cs42\" (UniqueName: \"kubernetes.io/projected/349c1264-8935-405e-a1bd-530a720bab81-kube-api-access-8cs42\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.875714 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-catalog-content\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.875775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-utilities\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.875813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cs42\" (UniqueName: \"kubernetes.io/projected/349c1264-8935-405e-a1bd-530a720bab81-kube-api-access-8cs42\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.876375 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-catalog-content\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.876449 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-utilities\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.898698 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cs42\" (UniqueName: \"kubernetes.io/projected/349c1264-8935-405e-a1bd-530a720bab81-kube-api-access-8cs42\") pod \"redhat-marketplace-956sh\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:44 crc kubenswrapper[4758]: I1203 17:53:44.998953 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:45 crc kubenswrapper[4758]: I1203 17:53:45.231749 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-956sh"] Dec 03 17:53:46 crc kubenswrapper[4758]: I1203 17:53:46.147289 4758 generic.go:334] "Generic (PLEG): container finished" podID="349c1264-8935-405e-a1bd-530a720bab81" containerID="e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992" exitCode=0 Dec 03 17:53:46 crc kubenswrapper[4758]: I1203 17:53:46.147348 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-956sh" event={"ID":"349c1264-8935-405e-a1bd-530a720bab81","Type":"ContainerDied","Data":"e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992"} Dec 03 17:53:46 crc kubenswrapper[4758]: I1203 17:53:46.147563 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-956sh" event={"ID":"349c1264-8935-405e-a1bd-530a720bab81","Type":"ContainerStarted","Data":"5d5cdef79618fc64b57dcefe636a3e48f7a2345cf29b6f4d86d454dfc040257a"} Dec 03 17:53:48 crc kubenswrapper[4758]: I1203 17:53:48.163827 4758 generic.go:334] "Generic (PLEG): container finished" podID="349c1264-8935-405e-a1bd-530a720bab81" containerID="12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9" exitCode=0 Dec 03 17:53:48 crc kubenswrapper[4758]: I1203 17:53:48.163929 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-956sh" event={"ID":"349c1264-8935-405e-a1bd-530a720bab81","Type":"ContainerDied","Data":"12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9"} Dec 03 17:53:49 crc kubenswrapper[4758]: I1203 17:53:49.173382 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-956sh" event={"ID":"349c1264-8935-405e-a1bd-530a720bab81","Type":"ContainerStarted","Data":"288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467"} Dec 03 17:53:49 crc kubenswrapper[4758]: I1203 17:53:49.188926 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-956sh" podStartSLOduration=2.517330104 podStartE2EDuration="5.188908399s" podCreationTimestamp="2025-12-03 17:53:44 +0000 UTC" firstStartedPulling="2025-12-03 17:53:46.149428036 +0000 UTC m=+3481.350804917" lastFinishedPulling="2025-12-03 17:53:48.821006351 +0000 UTC m=+3484.022383212" observedRunningTime="2025-12-03 17:53:49.188066547 +0000 UTC m=+3484.389443408" watchObservedRunningTime="2025-12-03 17:53:49.188908399 +0000 UTC m=+3484.390285260" Dec 03 17:53:55 crc kubenswrapper[4758]: I1203 17:53:54.999939 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:55 crc kubenswrapper[4758]: I1203 17:53:55.000973 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:55 crc kubenswrapper[4758]: I1203 17:53:55.044481 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:55 crc kubenswrapper[4758]: I1203 17:53:55.260200 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:55 crc kubenswrapper[4758]: I1203 17:53:55.303486 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-956sh"] Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.230323 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-956sh" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="registry-server" containerID="cri-o://288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467" gracePeriod=2 Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.609808 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.752160 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-catalog-content\") pod \"349c1264-8935-405e-a1bd-530a720bab81\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.752300 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cs42\" (UniqueName: \"kubernetes.io/projected/349c1264-8935-405e-a1bd-530a720bab81-kube-api-access-8cs42\") pod \"349c1264-8935-405e-a1bd-530a720bab81\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.752347 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-utilities\") pod \"349c1264-8935-405e-a1bd-530a720bab81\" (UID: \"349c1264-8935-405e-a1bd-530a720bab81\") " Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.753591 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-utilities" (OuterVolumeSpecName: "utilities") pod "349c1264-8935-405e-a1bd-530a720bab81" (UID: "349c1264-8935-405e-a1bd-530a720bab81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.760002 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/349c1264-8935-405e-a1bd-530a720bab81-kube-api-access-8cs42" (OuterVolumeSpecName: "kube-api-access-8cs42") pod "349c1264-8935-405e-a1bd-530a720bab81" (UID: "349c1264-8935-405e-a1bd-530a720bab81"). InnerVolumeSpecName "kube-api-access-8cs42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.771575 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "349c1264-8935-405e-a1bd-530a720bab81" (UID: "349c1264-8935-405e-a1bd-530a720bab81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.854570 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.854623 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349c1264-8935-405e-a1bd-530a720bab81-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:53:57 crc kubenswrapper[4758]: I1203 17:53:57.854638 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cs42\" (UniqueName: \"kubernetes.io/projected/349c1264-8935-405e-a1bd-530a720bab81-kube-api-access-8cs42\") on node \"crc\" DevicePath \"\"" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.239580 4758 generic.go:334] "Generic (PLEG): container finished" podID="349c1264-8935-405e-a1bd-530a720bab81" containerID="288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467" exitCode=0 Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.239630 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-956sh" event={"ID":"349c1264-8935-405e-a1bd-530a720bab81","Type":"ContainerDied","Data":"288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467"} Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.239645 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-956sh" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.239726 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-956sh" event={"ID":"349c1264-8935-405e-a1bd-530a720bab81","Type":"ContainerDied","Data":"5d5cdef79618fc64b57dcefe636a3e48f7a2345cf29b6f4d86d454dfc040257a"} Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.239753 4758 scope.go:117] "RemoveContainer" containerID="288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.277573 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-956sh"] Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.280310 4758 scope.go:117] "RemoveContainer" containerID="12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.286611 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-956sh"] Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.296957 4758 scope.go:117] "RemoveContainer" containerID="e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.319737 4758 scope.go:117] "RemoveContainer" containerID="288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467" Dec 03 17:53:58 crc kubenswrapper[4758]: E1203 17:53:58.320087 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467\": container with ID starting with 288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467 not found: ID does not exist" containerID="288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.320128 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467"} err="failed to get container status \"288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467\": rpc error: code = NotFound desc = could not find container \"288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467\": container with ID starting with 288c4993d4c5906de50c6848faba96650f061c297b3d44c9688a5e24d7b20467 not found: ID does not exist" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.320151 4758 scope.go:117] "RemoveContainer" containerID="12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9" Dec 03 17:53:58 crc kubenswrapper[4758]: E1203 17:53:58.320368 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9\": container with ID starting with 12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9 not found: ID does not exist" containerID="12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.320399 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9"} err="failed to get container status \"12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9\": rpc error: code = NotFound desc = could not find container \"12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9\": container with ID starting with 12c0c044619e55e2519cfefd87241fe76184ab8eaf4c8202e532b309d75904b9 not found: ID does not exist" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.320415 4758 scope.go:117] "RemoveContainer" containerID="e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992" Dec 03 17:53:58 crc kubenswrapper[4758]: E1203 17:53:58.320671 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992\": container with ID starting with e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992 not found: ID does not exist" containerID="e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992" Dec 03 17:53:58 crc kubenswrapper[4758]: I1203 17:53:58.320714 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992"} err="failed to get container status \"e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992\": rpc error: code = NotFound desc = could not find container \"e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992\": container with ID starting with e1556bd376034763499b7cf3d95a5a2792ca41b03bff1175df5f1c835689d992 not found: ID does not exist" Dec 03 17:53:59 crc kubenswrapper[4758]: I1203 17:53:59.126012 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="349c1264-8935-405e-a1bd-530a720bab81" path="/var/lib/kubelet/pods/349c1264-8935-405e-a1bd-530a720bab81/volumes" Dec 03 17:54:11 crc kubenswrapper[4758]: I1203 17:54:11.394233 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:54:11 crc kubenswrapper[4758]: I1203 17:54:11.394822 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.395058 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.395579 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.395620 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.396144 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ddb3a561d726a22607fe93840c810114243f0107c35ca00e535adbcce402682"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.396199 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://2ddb3a561d726a22607fe93840c810114243f0107c35ca00e535adbcce402682" gracePeriod=600 Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.556647 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="2ddb3a561d726a22607fe93840c810114243f0107c35ca00e535adbcce402682" exitCode=0 Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.556739 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"2ddb3a561d726a22607fe93840c810114243f0107c35ca00e535adbcce402682"} Dec 03 17:54:41 crc kubenswrapper[4758]: I1203 17:54:41.556818 4758 scope.go:117] "RemoveContainer" containerID="3caf9802af011a23b4540397d8a05cc598d5be44614bb07e8377fafe2e30ae16" Dec 03 17:54:42 crc kubenswrapper[4758]: I1203 17:54:42.568880 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218"} Dec 03 17:56:09 crc kubenswrapper[4758]: I1203 17:56:09.999098 4758 scope.go:117] "RemoveContainer" containerID="6eb6d3f2b5bbf81b0d78caf12a876da3c949eeda97366a189e1e6b00ed482aac" Dec 03 17:56:10 crc kubenswrapper[4758]: I1203 17:56:10.021506 4758 scope.go:117] "RemoveContainer" containerID="10ec827d058e7929f15bacb4fc96bbc5123cc5bf65676ba7a892e567b4dcd1e3" Dec 03 17:56:10 crc kubenswrapper[4758]: I1203 17:56:10.039950 4758 scope.go:117] "RemoveContainer" containerID="69ccae3ccc4b67df473ef681138055c2b5e92317c889a9f669dd558470991c15" Dec 03 17:56:41 crc kubenswrapper[4758]: I1203 17:56:41.395210 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:56:41 crc kubenswrapper[4758]: I1203 17:56:41.397706 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:57:11 crc kubenswrapper[4758]: I1203 17:57:11.394437 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:57:11 crc kubenswrapper[4758]: I1203 17:57:11.395185 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.394568 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.395962 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.396055 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.396752 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.396814 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" gracePeriod=600 Dec 03 17:57:41 crc kubenswrapper[4758]: E1203 17:57:41.542259 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.932463 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" exitCode=0 Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.932517 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218"} Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.932888 4758 scope.go:117] "RemoveContainer" containerID="2ddb3a561d726a22607fe93840c810114243f0107c35ca00e535adbcce402682" Dec 03 17:57:41 crc kubenswrapper[4758]: I1203 17:57:41.933464 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:57:41 crc kubenswrapper[4758]: E1203 17:57:41.933790 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:57:54 crc kubenswrapper[4758]: I1203 17:57:54.114165 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:57:54 crc kubenswrapper[4758]: E1203 17:57:54.114950 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:58:09 crc kubenswrapper[4758]: I1203 17:58:09.114700 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:58:09 crc kubenswrapper[4758]: E1203 17:58:09.115536 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:58:21 crc kubenswrapper[4758]: I1203 17:58:21.115353 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:58:21 crc kubenswrapper[4758]: E1203 17:58:21.116010 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:58:33 crc kubenswrapper[4758]: I1203 17:58:33.115137 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:58:33 crc kubenswrapper[4758]: E1203 17:58:33.116410 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:58:44 crc kubenswrapper[4758]: I1203 17:58:44.114815 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:58:44 crc kubenswrapper[4758]: E1203 17:58:44.115393 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:58:58 crc kubenswrapper[4758]: I1203 17:58:58.114276 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:58:58 crc kubenswrapper[4758]: E1203 17:58:58.115321 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:59:10 crc kubenswrapper[4758]: I1203 17:59:10.115331 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:59:10 crc kubenswrapper[4758]: E1203 17:59:10.116260 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:59:21 crc kubenswrapper[4758]: I1203 17:59:21.114357 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:59:21 crc kubenswrapper[4758]: E1203 17:59:21.114947 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:59:35 crc kubenswrapper[4758]: I1203 17:59:35.118015 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:59:35 crc kubenswrapper[4758]: E1203 17:59:35.118721 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 17:59:49 crc kubenswrapper[4758]: I1203 17:59:49.114411 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 17:59:49 crc kubenswrapper[4758]: E1203 17:59:49.115128 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.177541 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv"] Dec 03 18:00:00 crc kubenswrapper[4758]: E1203 18:00:00.178427 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="extract-utilities" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.178445 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="extract-utilities" Dec 03 18:00:00 crc kubenswrapper[4758]: E1203 18:00:00.178472 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="registry-server" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.178480 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="registry-server" Dec 03 18:00:00 crc kubenswrapper[4758]: E1203 18:00:00.178495 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="extract-content" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.178502 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="extract-content" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.178660 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="349c1264-8935-405e-a1bd-530a720bab81" containerName="registry-server" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.179224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.181646 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.181887 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.186573 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv"] Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.309333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5b93102-be10-46be-9cd0-c28fadccc29a-config-volume\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.309422 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5b93102-be10-46be-9cd0-c28fadccc29a-secret-volume\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.309475 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz4hc\" (UniqueName: \"kubernetes.io/projected/e5b93102-be10-46be-9cd0-c28fadccc29a-kube-api-access-xz4hc\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.410328 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5b93102-be10-46be-9cd0-c28fadccc29a-config-volume\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.410377 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5b93102-be10-46be-9cd0-c28fadccc29a-secret-volume\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.410412 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz4hc\" (UniqueName: \"kubernetes.io/projected/e5b93102-be10-46be-9cd0-c28fadccc29a-kube-api-access-xz4hc\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.411547 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5b93102-be10-46be-9cd0-c28fadccc29a-config-volume\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.416656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5b93102-be10-46be-9cd0-c28fadccc29a-secret-volume\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.429147 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz4hc\" (UniqueName: \"kubernetes.io/projected/e5b93102-be10-46be-9cd0-c28fadccc29a-kube-api-access-xz4hc\") pod \"collect-profiles-29413080-bcnvv\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.498349 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.909271 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv"] Dec 03 18:00:00 crc kubenswrapper[4758]: I1203 18:00:00.955268 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" event={"ID":"e5b93102-be10-46be-9cd0-c28fadccc29a","Type":"ContainerStarted","Data":"cdb610b5cd81808ed5b61b4a59282c9b8c76ee77e02e5c990f8e3da228137a2f"} Dec 03 18:00:01 crc kubenswrapper[4758]: I1203 18:00:01.964363 4758 generic.go:334] "Generic (PLEG): container finished" podID="e5b93102-be10-46be-9cd0-c28fadccc29a" containerID="e5f4dbed26a69d5eb8c679c05b88617effa0716dd60a8d9592de96c70722c0b3" exitCode=0 Dec 03 18:00:01 crc kubenswrapper[4758]: I1203 18:00:01.964425 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" event={"ID":"e5b93102-be10-46be-9cd0-c28fadccc29a","Type":"ContainerDied","Data":"e5f4dbed26a69d5eb8c679c05b88617effa0716dd60a8d9592de96c70722c0b3"} Dec 03 18:00:02 crc kubenswrapper[4758]: I1203 18:00:02.114444 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:00:02 crc kubenswrapper[4758]: E1203 18:00:02.114794 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.244358 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.353340 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5b93102-be10-46be-9cd0-c28fadccc29a-secret-volume\") pod \"e5b93102-be10-46be-9cd0-c28fadccc29a\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.353380 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5b93102-be10-46be-9cd0-c28fadccc29a-config-volume\") pod \"e5b93102-be10-46be-9cd0-c28fadccc29a\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.353555 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz4hc\" (UniqueName: \"kubernetes.io/projected/e5b93102-be10-46be-9cd0-c28fadccc29a-kube-api-access-xz4hc\") pod \"e5b93102-be10-46be-9cd0-c28fadccc29a\" (UID: \"e5b93102-be10-46be-9cd0-c28fadccc29a\") " Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.354308 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5b93102-be10-46be-9cd0-c28fadccc29a-config-volume" (OuterVolumeSpecName: "config-volume") pod "e5b93102-be10-46be-9cd0-c28fadccc29a" (UID: "e5b93102-be10-46be-9cd0-c28fadccc29a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.354484 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5b93102-be10-46be-9cd0-c28fadccc29a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.358944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5b93102-be10-46be-9cd0-c28fadccc29a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e5b93102-be10-46be-9cd0-c28fadccc29a" (UID: "e5b93102-be10-46be-9cd0-c28fadccc29a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.362521 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b93102-be10-46be-9cd0-c28fadccc29a-kube-api-access-xz4hc" (OuterVolumeSpecName: "kube-api-access-xz4hc") pod "e5b93102-be10-46be-9cd0-c28fadccc29a" (UID: "e5b93102-be10-46be-9cd0-c28fadccc29a"). InnerVolumeSpecName "kube-api-access-xz4hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.455720 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5b93102-be10-46be-9cd0-c28fadccc29a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.455772 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz4hc\" (UniqueName: \"kubernetes.io/projected/e5b93102-be10-46be-9cd0-c28fadccc29a-kube-api-access-xz4hc\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.982513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" event={"ID":"e5b93102-be10-46be-9cd0-c28fadccc29a","Type":"ContainerDied","Data":"cdb610b5cd81808ed5b61b4a59282c9b8c76ee77e02e5c990f8e3da228137a2f"} Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.982853 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdb610b5cd81808ed5b61b4a59282c9b8c76ee77e02e5c990f8e3da228137a2f" Dec 03 18:00:03 crc kubenswrapper[4758]: I1203 18:00:03.982566 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv" Dec 03 18:00:04 crc kubenswrapper[4758]: I1203 18:00:04.315037 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l"] Dec 03 18:00:04 crc kubenswrapper[4758]: I1203 18:00:04.320794 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-r4r7l"] Dec 03 18:00:05 crc kubenswrapper[4758]: I1203 18:00:05.123831 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f93e2eca-7a66-4f9a-9151-7767b96dcc47" path="/var/lib/kubelet/pods/f93e2eca-7a66-4f9a-9151-7767b96dcc47/volumes" Dec 03 18:00:10 crc kubenswrapper[4758]: I1203 18:00:10.133021 4758 scope.go:117] "RemoveContainer" containerID="8e5a54860065889dbf42cc1aef9277736db5b1da160cd695ee541ff59945aadc" Dec 03 18:00:13 crc kubenswrapper[4758]: I1203 18:00:13.114804 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:00:13 crc kubenswrapper[4758]: E1203 18:00:13.115745 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:00:28 crc kubenswrapper[4758]: I1203 18:00:28.114511 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:00:28 crc kubenswrapper[4758]: E1203 18:00:28.115269 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:00:41 crc kubenswrapper[4758]: I1203 18:00:41.114459 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:00:41 crc kubenswrapper[4758]: E1203 18:00:41.115529 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:00:55 crc kubenswrapper[4758]: I1203 18:00:55.130825 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:00:55 crc kubenswrapper[4758]: E1203 18:00:55.132431 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:01:10 crc kubenswrapper[4758]: I1203 18:01:10.114321 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:01:10 crc kubenswrapper[4758]: E1203 18:01:10.115121 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:01:21 crc kubenswrapper[4758]: I1203 18:01:21.115180 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:01:21 crc kubenswrapper[4758]: E1203 18:01:21.115942 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:01:35 crc kubenswrapper[4758]: I1203 18:01:35.119226 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:01:35 crc kubenswrapper[4758]: E1203 18:01:35.119791 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.360520 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8txzk"] Dec 03 18:01:41 crc kubenswrapper[4758]: E1203 18:01:41.361481 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b93102-be10-46be-9cd0-c28fadccc29a" containerName="collect-profiles" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.361499 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b93102-be10-46be-9cd0-c28fadccc29a" containerName="collect-profiles" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.362670 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b93102-be10-46be-9cd0-c28fadccc29a" containerName="collect-profiles" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.364005 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.375727 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8txzk"] Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.435328 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsvmw\" (UniqueName: \"kubernetes.io/projected/9997d72f-4477-4709-b760-ebd736d2d6cd-kube-api-access-zsvmw\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.435412 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-catalog-content\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.435466 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-utilities\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.536285 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsvmw\" (UniqueName: \"kubernetes.io/projected/9997d72f-4477-4709-b760-ebd736d2d6cd-kube-api-access-zsvmw\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.536395 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-catalog-content\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.536444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-utilities\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.537027 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-utilities\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.537175 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-catalog-content\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.555428 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsvmw\" (UniqueName: \"kubernetes.io/projected/9997d72f-4477-4709-b760-ebd736d2d6cd-kube-api-access-zsvmw\") pod \"certified-operators-8txzk\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:41 crc kubenswrapper[4758]: I1203 18:01:41.691394 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:42 crc kubenswrapper[4758]: I1203 18:01:42.290570 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8txzk"] Dec 03 18:01:42 crc kubenswrapper[4758]: I1203 18:01:42.721758 4758 generic.go:334] "Generic (PLEG): container finished" podID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerID="05200608f46aaa1cedced24baeff514565ffd0db8dd919001f1d260550ee3bce" exitCode=0 Dec 03 18:01:42 crc kubenswrapper[4758]: I1203 18:01:42.721796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerDied","Data":"05200608f46aaa1cedced24baeff514565ffd0db8dd919001f1d260550ee3bce"} Dec 03 18:01:42 crc kubenswrapper[4758]: I1203 18:01:42.722066 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerStarted","Data":"fdcbcbdd3d9a636884eb30a52e1599e19d0e798a8018a7711c5423b708c88adb"} Dec 03 18:01:42 crc kubenswrapper[4758]: I1203 18:01:42.723366 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:01:43 crc kubenswrapper[4758]: I1203 18:01:43.729607 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerStarted","Data":"6bdc9ea8c9ce48fca22a78e1002dabe25963d5aacdc031876152e1e6e68a8f53"} Dec 03 18:01:44 crc kubenswrapper[4758]: I1203 18:01:44.738977 4758 generic.go:334] "Generic (PLEG): container finished" podID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerID="6bdc9ea8c9ce48fca22a78e1002dabe25963d5aacdc031876152e1e6e68a8f53" exitCode=0 Dec 03 18:01:44 crc kubenswrapper[4758]: I1203 18:01:44.739048 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerDied","Data":"6bdc9ea8c9ce48fca22a78e1002dabe25963d5aacdc031876152e1e6e68a8f53"} Dec 03 18:01:45 crc kubenswrapper[4758]: I1203 18:01:45.748627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerStarted","Data":"9fb6570c8588b1ddea645397517ed28465da367c7cded0d2a2af4e19266a11ed"} Dec 03 18:01:46 crc kubenswrapper[4758]: I1203 18:01:46.115488 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:01:46 crc kubenswrapper[4758]: E1203 18:01:46.115844 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:01:51 crc kubenswrapper[4758]: I1203 18:01:51.692579 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:51 crc kubenswrapper[4758]: I1203 18:01:51.693051 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:51 crc kubenswrapper[4758]: I1203 18:01:51.733405 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:51 crc kubenswrapper[4758]: I1203 18:01:51.752923 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8txzk" podStartSLOduration=8.290338301 podStartE2EDuration="10.752901512s" podCreationTimestamp="2025-12-03 18:01:41 +0000 UTC" firstStartedPulling="2025-12-03 18:01:42.723140244 +0000 UTC m=+3957.924517105" lastFinishedPulling="2025-12-03 18:01:45.185703455 +0000 UTC m=+3960.387080316" observedRunningTime="2025-12-03 18:01:45.775220413 +0000 UTC m=+3960.976597294" watchObservedRunningTime="2025-12-03 18:01:51.752901512 +0000 UTC m=+3966.954278373" Dec 03 18:01:51 crc kubenswrapper[4758]: I1203 18:01:51.824193 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:51 crc kubenswrapper[4758]: I1203 18:01:51.967214 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8txzk"] Dec 03 18:01:53 crc kubenswrapper[4758]: I1203 18:01:53.798779 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8txzk" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="registry-server" containerID="cri-o://9fb6570c8588b1ddea645397517ed28465da367c7cded0d2a2af4e19266a11ed" gracePeriod=2 Dec 03 18:01:54 crc kubenswrapper[4758]: I1203 18:01:54.806506 4758 generic.go:334] "Generic (PLEG): container finished" podID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerID="9fb6570c8588b1ddea645397517ed28465da367c7cded0d2a2af4e19266a11ed" exitCode=0 Dec 03 18:01:54 crc kubenswrapper[4758]: I1203 18:01:54.806551 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerDied","Data":"9fb6570c8588b1ddea645397517ed28465da367c7cded0d2a2af4e19266a11ed"} Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.307774 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.421319 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-utilities\") pod \"9997d72f-4477-4709-b760-ebd736d2d6cd\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.421511 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsvmw\" (UniqueName: \"kubernetes.io/projected/9997d72f-4477-4709-b760-ebd736d2d6cd-kube-api-access-zsvmw\") pod \"9997d72f-4477-4709-b760-ebd736d2d6cd\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.421534 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-catalog-content\") pod \"9997d72f-4477-4709-b760-ebd736d2d6cd\" (UID: \"9997d72f-4477-4709-b760-ebd736d2d6cd\") " Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.422284 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-utilities" (OuterVolumeSpecName: "utilities") pod "9997d72f-4477-4709-b760-ebd736d2d6cd" (UID: "9997d72f-4477-4709-b760-ebd736d2d6cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.422767 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.427477 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9997d72f-4477-4709-b760-ebd736d2d6cd-kube-api-access-zsvmw" (OuterVolumeSpecName: "kube-api-access-zsvmw") pod "9997d72f-4477-4709-b760-ebd736d2d6cd" (UID: "9997d72f-4477-4709-b760-ebd736d2d6cd"). InnerVolumeSpecName "kube-api-access-zsvmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.484213 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9997d72f-4477-4709-b760-ebd736d2d6cd" (UID: "9997d72f-4477-4709-b760-ebd736d2d6cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.523927 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsvmw\" (UniqueName: \"kubernetes.io/projected/9997d72f-4477-4709-b760-ebd736d2d6cd-kube-api-access-zsvmw\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.523991 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9997d72f-4477-4709-b760-ebd736d2d6cd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.814471 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8txzk" event={"ID":"9997d72f-4477-4709-b760-ebd736d2d6cd","Type":"ContainerDied","Data":"fdcbcbdd3d9a636884eb30a52e1599e19d0e798a8018a7711c5423b708c88adb"} Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.815406 4758 scope.go:117] "RemoveContainer" containerID="9fb6570c8588b1ddea645397517ed28465da367c7cded0d2a2af4e19266a11ed" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.814748 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8txzk" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.832835 4758 scope.go:117] "RemoveContainer" containerID="6bdc9ea8c9ce48fca22a78e1002dabe25963d5aacdc031876152e1e6e68a8f53" Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.846122 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8txzk"] Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.856403 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8txzk"] Dec 03 18:01:55 crc kubenswrapper[4758]: I1203 18:01:55.870867 4758 scope.go:117] "RemoveContainer" containerID="05200608f46aaa1cedced24baeff514565ffd0db8dd919001f1d260550ee3bce" Dec 03 18:01:57 crc kubenswrapper[4758]: I1203 18:01:57.115408 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:01:57 crc kubenswrapper[4758]: E1203 18:01:57.115846 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:01:57 crc kubenswrapper[4758]: I1203 18:01:57.125290 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" path="/var/lib/kubelet/pods/9997d72f-4477-4709-b760-ebd736d2d6cd/volumes" Dec 03 18:02:11 crc kubenswrapper[4758]: I1203 18:02:11.114473 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:02:11 crc kubenswrapper[4758]: E1203 18:02:11.115362 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:02:22 crc kubenswrapper[4758]: I1203 18:02:22.114781 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:02:22 crc kubenswrapper[4758]: E1203 18:02:22.115615 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:02:34 crc kubenswrapper[4758]: I1203 18:02:34.115566 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:02:34 crc kubenswrapper[4758]: E1203 18:02:34.116421 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:02:46 crc kubenswrapper[4758]: I1203 18:02:46.114396 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:02:47 crc kubenswrapper[4758]: I1203 18:02:47.213746 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"58e77bed1561ca11a519a6059e71599484e99662605053f2d353fa9c07bdf2ce"} Dec 03 18:03:06 crc kubenswrapper[4758]: I1203 18:03:06.994261 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-svlpv"] Dec 03 18:03:06 crc kubenswrapper[4758]: E1203 18:03:06.995189 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="registry-server" Dec 03 18:03:06 crc kubenswrapper[4758]: I1203 18:03:06.995297 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="registry-server" Dec 03 18:03:06 crc kubenswrapper[4758]: E1203 18:03:06.995317 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="extract-utilities" Dec 03 18:03:06 crc kubenswrapper[4758]: I1203 18:03:06.995326 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="extract-utilities" Dec 03 18:03:06 crc kubenswrapper[4758]: E1203 18:03:06.995346 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="extract-content" Dec 03 18:03:06 crc kubenswrapper[4758]: I1203 18:03:06.995359 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="extract-content" Dec 03 18:03:06 crc kubenswrapper[4758]: I1203 18:03:06.995577 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9997d72f-4477-4709-b760-ebd736d2d6cd" containerName="registry-server" Dec 03 18:03:06 crc kubenswrapper[4758]: I1203 18:03:06.997157 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.018266 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-svlpv"] Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.141584 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mqw5\" (UniqueName: \"kubernetes.io/projected/bcd56212-265e-465c-8ea5-434184739e65-kube-api-access-9mqw5\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.141654 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-utilities\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.141706 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-catalog-content\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.243407 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-utilities\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.243458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-catalog-content\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.243603 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mqw5\" (UniqueName: \"kubernetes.io/projected/bcd56212-265e-465c-8ea5-434184739e65-kube-api-access-9mqw5\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.244895 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-utilities\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.245651 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-catalog-content\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.275113 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mqw5\" (UniqueName: \"kubernetes.io/projected/bcd56212-265e-465c-8ea5-434184739e65-kube-api-access-9mqw5\") pod \"redhat-operators-svlpv\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.316785 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:07 crc kubenswrapper[4758]: I1203 18:03:07.834717 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-svlpv"] Dec 03 18:03:08 crc kubenswrapper[4758]: I1203 18:03:08.379989 4758 generic.go:334] "Generic (PLEG): container finished" podID="bcd56212-265e-465c-8ea5-434184739e65" containerID="c31fa53a37723e86b080c80626c6869dccd907aa4dfdc4e1411f126215005c62" exitCode=0 Dec 03 18:03:08 crc kubenswrapper[4758]: I1203 18:03:08.380269 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerDied","Data":"c31fa53a37723e86b080c80626c6869dccd907aa4dfdc4e1411f126215005c62"} Dec 03 18:03:08 crc kubenswrapper[4758]: I1203 18:03:08.380301 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerStarted","Data":"2f42a7bff3db0e61b48faf5504859bceb19caa3d16242961d7099836de341ff4"} Dec 03 18:03:09 crc kubenswrapper[4758]: I1203 18:03:09.388527 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerStarted","Data":"d26554aaa5b18f91e34bdd8060629b625b3e9918fa87957e866ec2fe9933c154"} Dec 03 18:03:10 crc kubenswrapper[4758]: I1203 18:03:10.399547 4758 generic.go:334] "Generic (PLEG): container finished" podID="bcd56212-265e-465c-8ea5-434184739e65" containerID="d26554aaa5b18f91e34bdd8060629b625b3e9918fa87957e866ec2fe9933c154" exitCode=0 Dec 03 18:03:10 crc kubenswrapper[4758]: I1203 18:03:10.399597 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerDied","Data":"d26554aaa5b18f91e34bdd8060629b625b3e9918fa87957e866ec2fe9933c154"} Dec 03 18:03:11 crc kubenswrapper[4758]: I1203 18:03:11.409032 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerStarted","Data":"21bbe401e4967c8abdf287ef7887903170687443bddfef7726303286a4cb0ca1"} Dec 03 18:03:11 crc kubenswrapper[4758]: I1203 18:03:11.432125 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-svlpv" podStartSLOduration=2.940503964 podStartE2EDuration="5.432106377s" podCreationTimestamp="2025-12-03 18:03:06 +0000 UTC" firstStartedPulling="2025-12-03 18:03:08.382720931 +0000 UTC m=+4043.584097792" lastFinishedPulling="2025-12-03 18:03:10.874323344 +0000 UTC m=+4046.075700205" observedRunningTime="2025-12-03 18:03:11.426822345 +0000 UTC m=+4046.628199206" watchObservedRunningTime="2025-12-03 18:03:11.432106377 +0000 UTC m=+4046.633483238" Dec 03 18:03:17 crc kubenswrapper[4758]: I1203 18:03:17.317602 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:17 crc kubenswrapper[4758]: I1203 18:03:17.319340 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:17 crc kubenswrapper[4758]: I1203 18:03:17.371168 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:17 crc kubenswrapper[4758]: I1203 18:03:17.489698 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:17 crc kubenswrapper[4758]: I1203 18:03:17.609856 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-svlpv"] Dec 03 18:03:19 crc kubenswrapper[4758]: I1203 18:03:19.465398 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-svlpv" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="registry-server" containerID="cri-o://21bbe401e4967c8abdf287ef7887903170687443bddfef7726303286a4cb0ca1" gracePeriod=2 Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.493957 4758 generic.go:334] "Generic (PLEG): container finished" podID="bcd56212-265e-465c-8ea5-434184739e65" containerID="21bbe401e4967c8abdf287ef7887903170687443bddfef7726303286a4cb0ca1" exitCode=0 Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.494002 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerDied","Data":"21bbe401e4967c8abdf287ef7887903170687443bddfef7726303286a4cb0ca1"} Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.738063 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.781123 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-utilities\") pod \"bcd56212-265e-465c-8ea5-434184739e65\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.781187 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-catalog-content\") pod \"bcd56212-265e-465c-8ea5-434184739e65\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.781214 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mqw5\" (UniqueName: \"kubernetes.io/projected/bcd56212-265e-465c-8ea5-434184739e65-kube-api-access-9mqw5\") pod \"bcd56212-265e-465c-8ea5-434184739e65\" (UID: \"bcd56212-265e-465c-8ea5-434184739e65\") " Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.782102 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-utilities" (OuterVolumeSpecName: "utilities") pod "bcd56212-265e-465c-8ea5-434184739e65" (UID: "bcd56212-265e-465c-8ea5-434184739e65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.832137 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd56212-265e-465c-8ea5-434184739e65-kube-api-access-9mqw5" (OuterVolumeSpecName: "kube-api-access-9mqw5") pod "bcd56212-265e-465c-8ea5-434184739e65" (UID: "bcd56212-265e-465c-8ea5-434184739e65"). InnerVolumeSpecName "kube-api-access-9mqw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.882173 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.882200 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mqw5\" (UniqueName: \"kubernetes.io/projected/bcd56212-265e-465c-8ea5-434184739e65-kube-api-access-9mqw5\") on node \"crc\" DevicePath \"\"" Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.882783 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcd56212-265e-465c-8ea5-434184739e65" (UID: "bcd56212-265e-465c-8ea5-434184739e65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:03:22 crc kubenswrapper[4758]: I1203 18:03:22.983552 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd56212-265e-465c-8ea5-434184739e65-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:03:23 crc kubenswrapper[4758]: E1203 18:03:23.225299 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcd56212_265e_465c_8ea5_434184739e65.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:03:23 crc kubenswrapper[4758]: I1203 18:03:23.505535 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-svlpv" event={"ID":"bcd56212-265e-465c-8ea5-434184739e65","Type":"ContainerDied","Data":"2f42a7bff3db0e61b48faf5504859bceb19caa3d16242961d7099836de341ff4"} Dec 03 18:03:23 crc kubenswrapper[4758]: I1203 18:03:23.506000 4758 scope.go:117] "RemoveContainer" containerID="21bbe401e4967c8abdf287ef7887903170687443bddfef7726303286a4cb0ca1" Dec 03 18:03:23 crc kubenswrapper[4758]: I1203 18:03:23.505604 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-svlpv" Dec 03 18:03:23 crc kubenswrapper[4758]: I1203 18:03:23.530328 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-svlpv"] Dec 03 18:03:23 crc kubenswrapper[4758]: I1203 18:03:23.533538 4758 scope.go:117] "RemoveContainer" containerID="d26554aaa5b18f91e34bdd8060629b625b3e9918fa87957e866ec2fe9933c154" Dec 03 18:03:23 crc kubenswrapper[4758]: I1203 18:03:23.536705 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-svlpv"] Dec 03 18:03:24 crc kubenswrapper[4758]: I1203 18:03:24.042958 4758 scope.go:117] "RemoveContainer" containerID="c31fa53a37723e86b080c80626c6869dccd907aa4dfdc4e1411f126215005c62" Dec 03 18:03:25 crc kubenswrapper[4758]: I1203 18:03:25.124947 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcd56212-265e-465c-8ea5-434184739e65" path="/var/lib/kubelet/pods/bcd56212-265e-465c-8ea5-434184739e65/volumes" Dec 03 18:05:11 crc kubenswrapper[4758]: I1203 18:05:11.394426 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:05:11 crc kubenswrapper[4758]: I1203 18:05:11.395016 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.295877 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qgpqx"] Dec 03 18:05:22 crc kubenswrapper[4758]: E1203 18:05:22.301303 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="extract-content" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.301452 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="extract-content" Dec 03 18:05:22 crc kubenswrapper[4758]: E1203 18:05:22.301542 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="registry-server" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.301626 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="registry-server" Dec 03 18:05:22 crc kubenswrapper[4758]: E1203 18:05:22.301748 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="extract-utilities" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.301837 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="extract-utilities" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.302119 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd56212-265e-465c-8ea5-434184739e65" containerName="registry-server" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.303502 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.310924 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgpqx"] Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.350555 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh9dm\" (UniqueName: \"kubernetes.io/projected/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-kube-api-access-lh9dm\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.350883 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-utilities\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.351163 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-catalog-content\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.452892 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-catalog-content\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.452975 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh9dm\" (UniqueName: \"kubernetes.io/projected/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-kube-api-access-lh9dm\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.453007 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-utilities\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.453600 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-utilities\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.453918 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-catalog-content\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.477232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh9dm\" (UniqueName: \"kubernetes.io/projected/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-kube-api-access-lh9dm\") pod \"community-operators-qgpqx\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:22 crc kubenswrapper[4758]: I1203 18:05:22.638852 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:23 crc kubenswrapper[4758]: I1203 18:05:23.100661 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgpqx"] Dec 03 18:05:23 crc kubenswrapper[4758]: I1203 18:05:23.418383 4758 generic.go:334] "Generic (PLEG): container finished" podID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerID="1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723" exitCode=0 Dec 03 18:05:23 crc kubenswrapper[4758]: I1203 18:05:23.418436 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgpqx" event={"ID":"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf","Type":"ContainerDied","Data":"1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723"} Dec 03 18:05:23 crc kubenswrapper[4758]: I1203 18:05:23.418467 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgpqx" event={"ID":"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf","Type":"ContainerStarted","Data":"2cad7748c9af4d4c45d80df02b9934204faee93d418d88c53f2b94931818c87f"} Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.425541 4758 generic.go:334] "Generic (PLEG): container finished" podID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerID="470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32" exitCode=0 Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.425766 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgpqx" event={"ID":"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf","Type":"ContainerDied","Data":"470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32"} Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.685074 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qvcsc"] Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.687466 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.694902 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvcsc"] Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.796075 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-utilities\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.796205 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc9t7\" (UniqueName: \"kubernetes.io/projected/7d941ef0-97b7-4939-8819-d44ce4066173-kube-api-access-rc9t7\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.796285 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-catalog-content\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.897364 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-catalog-content\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.897426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-utilities\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.897478 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc9t7\" (UniqueName: \"kubernetes.io/projected/7d941ef0-97b7-4939-8819-d44ce4066173-kube-api-access-rc9t7\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.898055 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-catalog-content\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.898194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-utilities\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:24 crc kubenswrapper[4758]: I1203 18:05:24.919745 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc9t7\" (UniqueName: \"kubernetes.io/projected/7d941ef0-97b7-4939-8819-d44ce4066173-kube-api-access-rc9t7\") pod \"redhat-marketplace-qvcsc\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:25 crc kubenswrapper[4758]: I1203 18:05:25.004731 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:25 crc kubenswrapper[4758]: I1203 18:05:25.436143 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgpqx" event={"ID":"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf","Type":"ContainerStarted","Data":"bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057"} Dec 03 18:05:25 crc kubenswrapper[4758]: I1203 18:05:25.460016 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qgpqx" podStartSLOduration=2.050853833 podStartE2EDuration="3.460000723s" podCreationTimestamp="2025-12-03 18:05:22 +0000 UTC" firstStartedPulling="2025-12-03 18:05:23.419772844 +0000 UTC m=+4178.621149705" lastFinishedPulling="2025-12-03 18:05:24.828919734 +0000 UTC m=+4180.030296595" observedRunningTime="2025-12-03 18:05:25.452949743 +0000 UTC m=+4180.654326604" watchObservedRunningTime="2025-12-03 18:05:25.460000723 +0000 UTC m=+4180.661377584" Dec 03 18:05:25 crc kubenswrapper[4758]: I1203 18:05:25.486744 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvcsc"] Dec 03 18:05:26 crc kubenswrapper[4758]: I1203 18:05:26.446354 4758 generic.go:334] "Generic (PLEG): container finished" podID="7d941ef0-97b7-4939-8819-d44ce4066173" containerID="326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac" exitCode=0 Dec 03 18:05:26 crc kubenswrapper[4758]: I1203 18:05:26.446428 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerDied","Data":"326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac"} Dec 03 18:05:26 crc kubenswrapper[4758]: I1203 18:05:26.446732 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerStarted","Data":"5da40e50d0b07a50dc3e949862be0923b8c81fc7f22761171dbb43b597ec8350"} Dec 03 18:05:27 crc kubenswrapper[4758]: I1203 18:05:27.455414 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerStarted","Data":"5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1"} Dec 03 18:05:28 crc kubenswrapper[4758]: I1203 18:05:28.473043 4758 generic.go:334] "Generic (PLEG): container finished" podID="7d941ef0-97b7-4939-8819-d44ce4066173" containerID="5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1" exitCode=0 Dec 03 18:05:28 crc kubenswrapper[4758]: I1203 18:05:28.473076 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerDied","Data":"5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1"} Dec 03 18:05:29 crc kubenswrapper[4758]: I1203 18:05:29.482374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerStarted","Data":"abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251"} Dec 03 18:05:29 crc kubenswrapper[4758]: I1203 18:05:29.512283 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qvcsc" podStartSLOduration=3.07874268 podStartE2EDuration="5.512247518s" podCreationTimestamp="2025-12-03 18:05:24 +0000 UTC" firstStartedPulling="2025-12-03 18:05:26.449147321 +0000 UTC m=+4181.650524192" lastFinishedPulling="2025-12-03 18:05:28.882652129 +0000 UTC m=+4184.084029030" observedRunningTime="2025-12-03 18:05:29.503705408 +0000 UTC m=+4184.705082269" watchObservedRunningTime="2025-12-03 18:05:29.512247518 +0000 UTC m=+4184.713624399" Dec 03 18:05:32 crc kubenswrapper[4758]: I1203 18:05:32.640280 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:32 crc kubenswrapper[4758]: I1203 18:05:32.640603 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:32 crc kubenswrapper[4758]: I1203 18:05:32.683263 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:33 crc kubenswrapper[4758]: I1203 18:05:33.552223 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:33 crc kubenswrapper[4758]: I1203 18:05:33.595348 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgpqx"] Dec 03 18:05:35 crc kubenswrapper[4758]: I1203 18:05:35.005318 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:35 crc kubenswrapper[4758]: I1203 18:05:35.005750 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:35 crc kubenswrapper[4758]: I1203 18:05:35.058925 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:35 crc kubenswrapper[4758]: I1203 18:05:35.521985 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qgpqx" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="registry-server" containerID="cri-o://bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057" gracePeriod=2 Dec 03 18:05:35 crc kubenswrapper[4758]: I1203 18:05:35.561938 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:35 crc kubenswrapper[4758]: I1203 18:05:35.882075 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvcsc"] Dec 03 18:05:36 crc kubenswrapper[4758]: I1203 18:05:36.986099 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.074590 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-utilities\") pod \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.074654 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-catalog-content\") pod \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.074796 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh9dm\" (UniqueName: \"kubernetes.io/projected/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-kube-api-access-lh9dm\") pod \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\" (UID: \"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf\") " Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.075586 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-utilities" (OuterVolumeSpecName: "utilities") pod "bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" (UID: "bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.079524 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-kube-api-access-lh9dm" (OuterVolumeSpecName: "kube-api-access-lh9dm") pod "bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" (UID: "bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf"). InnerVolumeSpecName "kube-api-access-lh9dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.129402 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" (UID: "bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.177198 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.177234 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.177249 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh9dm\" (UniqueName: \"kubernetes.io/projected/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf-kube-api-access-lh9dm\") on node \"crc\" DevicePath \"\"" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.535342 4758 generic.go:334] "Generic (PLEG): container finished" podID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerID="bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057" exitCode=0 Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.535408 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgpqx" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.535449 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgpqx" event={"ID":"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf","Type":"ContainerDied","Data":"bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057"} Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.536045 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgpqx" event={"ID":"bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf","Type":"ContainerDied","Data":"2cad7748c9af4d4c45d80df02b9934204faee93d418d88c53f2b94931818c87f"} Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.536092 4758 scope.go:117] "RemoveContainer" containerID="bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.536472 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qvcsc" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="registry-server" containerID="cri-o://abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251" gracePeriod=2 Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.556844 4758 scope.go:117] "RemoveContainer" containerID="470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.584879 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgpqx"] Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.590371 4758 scope.go:117] "RemoveContainer" containerID="1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.591982 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qgpqx"] Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.697623 4758 scope.go:117] "RemoveContainer" containerID="bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057" Dec 03 18:05:37 crc kubenswrapper[4758]: E1203 18:05:37.698241 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057\": container with ID starting with bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057 not found: ID does not exist" containerID="bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.698284 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057"} err="failed to get container status \"bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057\": rpc error: code = NotFound desc = could not find container \"bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057\": container with ID starting with bd8567ecf225057ea5383b6183a7bbba6ce570df9eb5b06868f4fd8bc1a5b057 not found: ID does not exist" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.698314 4758 scope.go:117] "RemoveContainer" containerID="470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32" Dec 03 18:05:37 crc kubenswrapper[4758]: E1203 18:05:37.698686 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32\": container with ID starting with 470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32 not found: ID does not exist" containerID="470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.698800 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32"} err="failed to get container status \"470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32\": rpc error: code = NotFound desc = could not find container \"470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32\": container with ID starting with 470bdd611fbc632d0322a963788310f09b340099f0e248e9424ebb5e069eeb32 not found: ID does not exist" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.698879 4758 scope.go:117] "RemoveContainer" containerID="1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723" Dec 03 18:05:37 crc kubenswrapper[4758]: E1203 18:05:37.699255 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723\": container with ID starting with 1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723 not found: ID does not exist" containerID="1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723" Dec 03 18:05:37 crc kubenswrapper[4758]: I1203 18:05:37.699292 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723"} err="failed to get container status \"1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723\": rpc error: code = NotFound desc = could not find container \"1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723\": container with ID starting with 1b47abb70c12fb91b3cafbf6606f5f9f7e10ea7bafe5169def4b553e364b0723 not found: ID does not exist" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.209071 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.290967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-utilities\") pod \"7d941ef0-97b7-4939-8819-d44ce4066173\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.291070 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-catalog-content\") pod \"7d941ef0-97b7-4939-8819-d44ce4066173\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.291110 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc9t7\" (UniqueName: \"kubernetes.io/projected/7d941ef0-97b7-4939-8819-d44ce4066173-kube-api-access-rc9t7\") pod \"7d941ef0-97b7-4939-8819-d44ce4066173\" (UID: \"7d941ef0-97b7-4939-8819-d44ce4066173\") " Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.292702 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-utilities" (OuterVolumeSpecName: "utilities") pod "7d941ef0-97b7-4939-8819-d44ce4066173" (UID: "7d941ef0-97b7-4939-8819-d44ce4066173"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.295996 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d941ef0-97b7-4939-8819-d44ce4066173-kube-api-access-rc9t7" (OuterVolumeSpecName: "kube-api-access-rc9t7") pod "7d941ef0-97b7-4939-8819-d44ce4066173" (UID: "7d941ef0-97b7-4939-8819-d44ce4066173"). InnerVolumeSpecName "kube-api-access-rc9t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.309388 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d941ef0-97b7-4939-8819-d44ce4066173" (UID: "7d941ef0-97b7-4939-8819-d44ce4066173"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.393390 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.393427 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d941ef0-97b7-4939-8819-d44ce4066173-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.393443 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc9t7\" (UniqueName: \"kubernetes.io/projected/7d941ef0-97b7-4939-8819-d44ce4066173-kube-api-access-rc9t7\") on node \"crc\" DevicePath \"\"" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.546562 4758 generic.go:334] "Generic (PLEG): container finished" podID="7d941ef0-97b7-4939-8819-d44ce4066173" containerID="abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251" exitCode=0 Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.546604 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerDied","Data":"abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251"} Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.546631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvcsc" event={"ID":"7d941ef0-97b7-4939-8819-d44ce4066173","Type":"ContainerDied","Data":"5da40e50d0b07a50dc3e949862be0923b8c81fc7f22761171dbb43b597ec8350"} Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.546648 4758 scope.go:117] "RemoveContainer" containerID="abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.546717 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvcsc" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.565952 4758 scope.go:117] "RemoveContainer" containerID="5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.580810 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvcsc"] Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.585901 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvcsc"] Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.604176 4758 scope.go:117] "RemoveContainer" containerID="326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.622904 4758 scope.go:117] "RemoveContainer" containerID="abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251" Dec 03 18:05:38 crc kubenswrapper[4758]: E1203 18:05:38.623347 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251\": container with ID starting with abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251 not found: ID does not exist" containerID="abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.623373 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251"} err="failed to get container status \"abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251\": rpc error: code = NotFound desc = could not find container \"abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251\": container with ID starting with abe1520ce08b6b909a2e351f4948df15edae82f65cd3e5944eb4f29542323251 not found: ID does not exist" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.623394 4758 scope.go:117] "RemoveContainer" containerID="5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1" Dec 03 18:05:38 crc kubenswrapper[4758]: E1203 18:05:38.623815 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1\": container with ID starting with 5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1 not found: ID does not exist" containerID="5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.623841 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1"} err="failed to get container status \"5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1\": rpc error: code = NotFound desc = could not find container \"5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1\": container with ID starting with 5d55e00e19f1bd0e2cc9ce76b5aab70a4f17187cba0474cc67896c374e3848a1 not found: ID does not exist" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.623879 4758 scope.go:117] "RemoveContainer" containerID="326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac" Dec 03 18:05:38 crc kubenswrapper[4758]: E1203 18:05:38.624340 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac\": container with ID starting with 326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac not found: ID does not exist" containerID="326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac" Dec 03 18:05:38 crc kubenswrapper[4758]: I1203 18:05:38.624395 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac"} err="failed to get container status \"326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac\": rpc error: code = NotFound desc = could not find container \"326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac\": container with ID starting with 326b7f9f55c8ece91d13ea31493ab3f078ac21299be8d2de2f5ebdda4e6f36ac not found: ID does not exist" Dec 03 18:05:39 crc kubenswrapper[4758]: I1203 18:05:39.123030 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" path="/var/lib/kubelet/pods/7d941ef0-97b7-4939-8819-d44ce4066173/volumes" Dec 03 18:05:39 crc kubenswrapper[4758]: I1203 18:05:39.124147 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" path="/var/lib/kubelet/pods/bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf/volumes" Dec 03 18:05:41 crc kubenswrapper[4758]: I1203 18:05:41.394950 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:05:41 crc kubenswrapper[4758]: I1203 18:05:41.395006 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.394856 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.395447 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.395501 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.396220 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58e77bed1561ca11a519a6059e71599484e99662605053f2d353fa9c07bdf2ce"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.396276 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://58e77bed1561ca11a519a6059e71599484e99662605053f2d353fa9c07bdf2ce" gracePeriod=600 Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.797435 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="58e77bed1561ca11a519a6059e71599484e99662605053f2d353fa9c07bdf2ce" exitCode=0 Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.797521 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"58e77bed1561ca11a519a6059e71599484e99662605053f2d353fa9c07bdf2ce"} Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.798092 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949"} Dec 03 18:06:11 crc kubenswrapper[4758]: I1203 18:06:11.798196 4758 scope.go:117] "RemoveContainer" containerID="4e8817b1e9a508bc5e7a7b2ea9822229c0394320b2c40ae9a1df36628ff4f218" Dec 03 18:08:11 crc kubenswrapper[4758]: I1203 18:08:11.394764 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:08:11 crc kubenswrapper[4758]: I1203 18:08:11.395366 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:08:41 crc kubenswrapper[4758]: I1203 18:08:41.395291 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:08:41 crc kubenswrapper[4758]: I1203 18:08:41.396052 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:09:11 crc kubenswrapper[4758]: I1203 18:09:11.394431 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:09:11 crc kubenswrapper[4758]: I1203 18:09:11.395086 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:09:11 crc kubenswrapper[4758]: I1203 18:09:11.395147 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:09:11 crc kubenswrapper[4758]: I1203 18:09:11.396030 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:09:11 crc kubenswrapper[4758]: I1203 18:09:11.396129 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" gracePeriod=600 Dec 03 18:09:11 crc kubenswrapper[4758]: E1203 18:09:11.525735 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:09:12 crc kubenswrapper[4758]: I1203 18:09:12.167792 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" exitCode=0 Dec 03 18:09:12 crc kubenswrapper[4758]: I1203 18:09:12.167843 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949"} Dec 03 18:09:12 crc kubenswrapper[4758]: I1203 18:09:12.167879 4758 scope.go:117] "RemoveContainer" containerID="58e77bed1561ca11a519a6059e71599484e99662605053f2d353fa9c07bdf2ce" Dec 03 18:09:12 crc kubenswrapper[4758]: I1203 18:09:12.168486 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:09:12 crc kubenswrapper[4758]: E1203 18:09:12.168826 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:09:24 crc kubenswrapper[4758]: I1203 18:09:24.114486 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:09:24 crc kubenswrapper[4758]: E1203 18:09:24.115245 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:09:39 crc kubenswrapper[4758]: I1203 18:09:39.114332 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:09:39 crc kubenswrapper[4758]: E1203 18:09:39.114953 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:09:53 crc kubenswrapper[4758]: I1203 18:09:53.114500 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:09:53 crc kubenswrapper[4758]: E1203 18:09:53.115181 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:10:04 crc kubenswrapper[4758]: I1203 18:10:04.115111 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:10:04 crc kubenswrapper[4758]: E1203 18:10:04.115887 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:10:18 crc kubenswrapper[4758]: I1203 18:10:18.113931 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:10:18 crc kubenswrapper[4758]: E1203 18:10:18.114637 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:10:33 crc kubenswrapper[4758]: I1203 18:10:33.114438 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:10:33 crc kubenswrapper[4758]: E1203 18:10:33.115244 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:10:44 crc kubenswrapper[4758]: I1203 18:10:44.115098 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:10:44 crc kubenswrapper[4758]: E1203 18:10:44.116234 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:10:56 crc kubenswrapper[4758]: I1203 18:10:56.113829 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:10:56 crc kubenswrapper[4758]: E1203 18:10:56.114837 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:11:09 crc kubenswrapper[4758]: I1203 18:11:09.114454 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:11:09 crc kubenswrapper[4758]: E1203 18:11:09.115117 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:11:23 crc kubenswrapper[4758]: I1203 18:11:23.114904 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:11:23 crc kubenswrapper[4758]: E1203 18:11:23.115758 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:11:35 crc kubenswrapper[4758]: I1203 18:11:35.119602 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:11:35 crc kubenswrapper[4758]: E1203 18:11:35.120808 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:11:49 crc kubenswrapper[4758]: I1203 18:11:49.115252 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:11:49 crc kubenswrapper[4758]: E1203 18:11:49.118285 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.026666 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xrn6f"] Dec 03 18:11:54 crc kubenswrapper[4758]: E1203 18:11:54.027788 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="extract-utilities" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.027815 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="extract-utilities" Dec 03 18:11:54 crc kubenswrapper[4758]: E1203 18:11:54.027836 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.027848 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4758]: E1203 18:11:54.027875 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.027888 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4758]: E1203 18:11:54.027905 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="extract-utilities" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.027917 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="extract-utilities" Dec 03 18:11:54 crc kubenswrapper[4758]: E1203 18:11:54.027939 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="extract-content" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.027951 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="extract-content" Dec 03 18:11:54 crc kubenswrapper[4758]: E1203 18:11:54.027986 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="extract-content" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.028000 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="extract-content" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.028247 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2b48fa-12de-4a6f-9ec2-66c6bbd32abf" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.028274 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d941ef0-97b7-4939-8819-d44ce4066173" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.031490 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.042648 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xrn6f"] Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.066936 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrdxq\" (UniqueName: \"kubernetes.io/projected/4051e00c-31de-48d3-9b13-21086d61da1e-kube-api-access-qrdxq\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.067119 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-utilities\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.067518 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-catalog-content\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.168786 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-catalog-content\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.169209 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrdxq\" (UniqueName: \"kubernetes.io/projected/4051e00c-31de-48d3-9b13-21086d61da1e-kube-api-access-qrdxq\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.169257 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-utilities\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.169573 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-catalog-content\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.169770 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-utilities\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.192962 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrdxq\" (UniqueName: \"kubernetes.io/projected/4051e00c-31de-48d3-9b13-21086d61da1e-kube-api-access-qrdxq\") pod \"certified-operators-xrn6f\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.375404 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:11:54 crc kubenswrapper[4758]: I1203 18:11:54.857294 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xrn6f"] Dec 03 18:11:55 crc kubenswrapper[4758]: I1203 18:11:55.457578 4758 generic.go:334] "Generic (PLEG): container finished" podID="4051e00c-31de-48d3-9b13-21086d61da1e" containerID="2b2d48ead0e0004540aab6b348743ad1d1ba996ce200c5e8f267dd5329ab9f3e" exitCode=0 Dec 03 18:11:55 crc kubenswrapper[4758]: I1203 18:11:55.457652 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerDied","Data":"2b2d48ead0e0004540aab6b348743ad1d1ba996ce200c5e8f267dd5329ab9f3e"} Dec 03 18:11:55 crc kubenswrapper[4758]: I1203 18:11:55.457908 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerStarted","Data":"49a57ba178727b7a64bb24e222aed30c17bbed1ebe9bfd8f8d041b3a36c4b136"} Dec 03 18:11:55 crc kubenswrapper[4758]: I1203 18:11:55.459451 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:11:56 crc kubenswrapper[4758]: I1203 18:11:56.470313 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerStarted","Data":"24793c036756c6a71222f68627ce21eab13fc0b90b3aee62b860457d920f8f88"} Dec 03 18:11:57 crc kubenswrapper[4758]: I1203 18:11:57.478927 4758 generic.go:334] "Generic (PLEG): container finished" podID="4051e00c-31de-48d3-9b13-21086d61da1e" containerID="24793c036756c6a71222f68627ce21eab13fc0b90b3aee62b860457d920f8f88" exitCode=0 Dec 03 18:11:57 crc kubenswrapper[4758]: I1203 18:11:57.478993 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerDied","Data":"24793c036756c6a71222f68627ce21eab13fc0b90b3aee62b860457d920f8f88"} Dec 03 18:11:58 crc kubenswrapper[4758]: I1203 18:11:58.492027 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerStarted","Data":"f2579bfac8076ca1000b349b47228bbab3ca116450185cbe616e2d2c2fae70e0"} Dec 03 18:11:58 crc kubenswrapper[4758]: I1203 18:11:58.520803 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xrn6f" podStartSLOduration=3.102931707 podStartE2EDuration="5.520786247s" podCreationTimestamp="2025-12-03 18:11:53 +0000 UTC" firstStartedPulling="2025-12-03 18:11:55.459189338 +0000 UTC m=+4570.660566199" lastFinishedPulling="2025-12-03 18:11:57.877043878 +0000 UTC m=+4573.078420739" observedRunningTime="2025-12-03 18:11:58.516353748 +0000 UTC m=+4573.717730609" watchObservedRunningTime="2025-12-03 18:11:58.520786247 +0000 UTC m=+4573.722163108" Dec 03 18:12:01 crc kubenswrapper[4758]: I1203 18:12:01.114193 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:12:01 crc kubenswrapper[4758]: E1203 18:12:01.114432 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:12:04 crc kubenswrapper[4758]: I1203 18:12:04.375576 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:12:04 crc kubenswrapper[4758]: I1203 18:12:04.375934 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:12:04 crc kubenswrapper[4758]: I1203 18:12:04.427586 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:12:04 crc kubenswrapper[4758]: I1203 18:12:04.568304 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:12:04 crc kubenswrapper[4758]: I1203 18:12:04.656106 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xrn6f"] Dec 03 18:12:06 crc kubenswrapper[4758]: I1203 18:12:06.544112 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xrn6f" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="registry-server" containerID="cri-o://f2579bfac8076ca1000b349b47228bbab3ca116450185cbe616e2d2c2fae70e0" gracePeriod=2 Dec 03 18:12:07 crc kubenswrapper[4758]: I1203 18:12:07.552100 4758 generic.go:334] "Generic (PLEG): container finished" podID="4051e00c-31de-48d3-9b13-21086d61da1e" containerID="f2579bfac8076ca1000b349b47228bbab3ca116450185cbe616e2d2c2fae70e0" exitCode=0 Dec 03 18:12:07 crc kubenswrapper[4758]: I1203 18:12:07.552147 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerDied","Data":"f2579bfac8076ca1000b349b47228bbab3ca116450185cbe616e2d2c2fae70e0"} Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.017113 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.073514 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrdxq\" (UniqueName: \"kubernetes.io/projected/4051e00c-31de-48d3-9b13-21086d61da1e-kube-api-access-qrdxq\") pod \"4051e00c-31de-48d3-9b13-21086d61da1e\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.073698 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-catalog-content\") pod \"4051e00c-31de-48d3-9b13-21086d61da1e\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.073801 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-utilities\") pod \"4051e00c-31de-48d3-9b13-21086d61da1e\" (UID: \"4051e00c-31de-48d3-9b13-21086d61da1e\") " Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.074643 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-utilities" (OuterVolumeSpecName: "utilities") pod "4051e00c-31de-48d3-9b13-21086d61da1e" (UID: "4051e00c-31de-48d3-9b13-21086d61da1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.079983 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4051e00c-31de-48d3-9b13-21086d61da1e-kube-api-access-qrdxq" (OuterVolumeSpecName: "kube-api-access-qrdxq") pod "4051e00c-31de-48d3-9b13-21086d61da1e" (UID: "4051e00c-31de-48d3-9b13-21086d61da1e"). InnerVolumeSpecName "kube-api-access-qrdxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.122028 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4051e00c-31de-48d3-9b13-21086d61da1e" (UID: "4051e00c-31de-48d3-9b13-21086d61da1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.175429 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.175474 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4051e00c-31de-48d3-9b13-21086d61da1e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.175490 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrdxq\" (UniqueName: \"kubernetes.io/projected/4051e00c-31de-48d3-9b13-21086d61da1e-kube-api-access-qrdxq\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.562796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrn6f" event={"ID":"4051e00c-31de-48d3-9b13-21086d61da1e","Type":"ContainerDied","Data":"49a57ba178727b7a64bb24e222aed30c17bbed1ebe9bfd8f8d041b3a36c4b136"} Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.562878 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrn6f" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.563181 4758 scope.go:117] "RemoveContainer" containerID="f2579bfac8076ca1000b349b47228bbab3ca116450185cbe616e2d2c2fae70e0" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.580731 4758 scope.go:117] "RemoveContainer" containerID="24793c036756c6a71222f68627ce21eab13fc0b90b3aee62b860457d920f8f88" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.606503 4758 scope.go:117] "RemoveContainer" containerID="2b2d48ead0e0004540aab6b348743ad1d1ba996ce200c5e8f267dd5329ab9f3e" Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.612769 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xrn6f"] Dec 03 18:12:08 crc kubenswrapper[4758]: I1203 18:12:08.621810 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xrn6f"] Dec 03 18:12:09 crc kubenswrapper[4758]: I1203 18:12:09.125356 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" path="/var/lib/kubelet/pods/4051e00c-31de-48d3-9b13-21086d61da1e/volumes" Dec 03 18:12:16 crc kubenswrapper[4758]: I1203 18:12:16.115342 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:12:16 crc kubenswrapper[4758]: E1203 18:12:16.116242 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:12:28 crc kubenswrapper[4758]: I1203 18:12:28.114106 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:12:28 crc kubenswrapper[4758]: E1203 18:12:28.115128 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:12:41 crc kubenswrapper[4758]: I1203 18:12:41.114321 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:12:41 crc kubenswrapper[4758]: E1203 18:12:41.115215 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:12:55 crc kubenswrapper[4758]: I1203 18:12:55.118451 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:12:55 crc kubenswrapper[4758]: E1203 18:12:55.119184 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:13:07 crc kubenswrapper[4758]: I1203 18:13:07.115032 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:13:07 crc kubenswrapper[4758]: E1203 18:13:07.115939 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:13:18 crc kubenswrapper[4758]: I1203 18:13:18.114797 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:13:18 crc kubenswrapper[4758]: E1203 18:13:18.115896 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:13:30 crc kubenswrapper[4758]: I1203 18:13:30.114240 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:13:30 crc kubenswrapper[4758]: E1203 18:13:30.114991 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:13:44 crc kubenswrapper[4758]: I1203 18:13:44.114349 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:13:44 crc kubenswrapper[4758]: E1203 18:13:44.115016 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:13:56 crc kubenswrapper[4758]: I1203 18:13:56.113928 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:13:56 crc kubenswrapper[4758]: E1203 18:13:56.114885 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.304567 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hbx7w"] Dec 03 18:13:57 crc kubenswrapper[4758]: E1203 18:13:57.305256 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="registry-server" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.305272 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="registry-server" Dec 03 18:13:57 crc kubenswrapper[4758]: E1203 18:13:57.305294 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="extract-utilities" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.305302 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="extract-utilities" Dec 03 18:13:57 crc kubenswrapper[4758]: E1203 18:13:57.305311 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="extract-content" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.305319 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="extract-content" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.305480 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4051e00c-31de-48d3-9b13-21086d61da1e" containerName="registry-server" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.307086 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.321612 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbx7w"] Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.448945 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-utilities\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.448997 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-catalog-content\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.449041 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw4n8\" (UniqueName: \"kubernetes.io/projected/d38dca61-3c59-47c9-9605-3cf815add7bf-kube-api-access-xw4n8\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.550477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-utilities\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.550533 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-catalog-content\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.550585 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw4n8\" (UniqueName: \"kubernetes.io/projected/d38dca61-3c59-47c9-9605-3cf815add7bf-kube-api-access-xw4n8\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.551316 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-catalog-content\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.551589 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-utilities\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.578252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw4n8\" (UniqueName: \"kubernetes.io/projected/d38dca61-3c59-47c9-9605-3cf815add7bf-kube-api-access-xw4n8\") pod \"redhat-operators-hbx7w\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:57 crc kubenswrapper[4758]: I1203 18:13:57.624971 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:13:58 crc kubenswrapper[4758]: I1203 18:13:58.061260 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbx7w"] Dec 03 18:13:58 crc kubenswrapper[4758]: I1203 18:13:58.423902 4758 generic.go:334] "Generic (PLEG): container finished" podID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerID="7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2" exitCode=0 Dec 03 18:13:58 crc kubenswrapper[4758]: I1203 18:13:58.423996 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerDied","Data":"7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2"} Dec 03 18:13:58 crc kubenswrapper[4758]: I1203 18:13:58.424166 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerStarted","Data":"5d87fe81bd79b8b00ae4b2f623a23ed6d0ad12a97dd335bd98aa40c0fe46303b"} Dec 03 18:13:59 crc kubenswrapper[4758]: I1203 18:13:59.447570 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerStarted","Data":"6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765"} Dec 03 18:14:00 crc kubenswrapper[4758]: I1203 18:14:00.461131 4758 generic.go:334] "Generic (PLEG): container finished" podID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerID="6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765" exitCode=0 Dec 03 18:14:00 crc kubenswrapper[4758]: I1203 18:14:00.461189 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerDied","Data":"6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765"} Dec 03 18:14:01 crc kubenswrapper[4758]: I1203 18:14:01.473541 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerStarted","Data":"6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3"} Dec 03 18:14:01 crc kubenswrapper[4758]: I1203 18:14:01.498874 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hbx7w" podStartSLOduration=2.057442183 podStartE2EDuration="4.498833354s" podCreationTimestamp="2025-12-03 18:13:57 +0000 UTC" firstStartedPulling="2025-12-03 18:13:58.425735886 +0000 UTC m=+4693.627112747" lastFinishedPulling="2025-12-03 18:14:00.867127057 +0000 UTC m=+4696.068503918" observedRunningTime="2025-12-03 18:14:01.493784958 +0000 UTC m=+4696.695161829" watchObservedRunningTime="2025-12-03 18:14:01.498833354 +0000 UTC m=+4696.700210215" Dec 03 18:14:07 crc kubenswrapper[4758]: I1203 18:14:07.625144 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:14:07 crc kubenswrapper[4758]: I1203 18:14:07.625463 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:14:07 crc kubenswrapper[4758]: I1203 18:14:07.676912 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:14:08 crc kubenswrapper[4758]: I1203 18:14:08.563060 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:14:08 crc kubenswrapper[4758]: I1203 18:14:08.613296 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbx7w"] Dec 03 18:14:10 crc kubenswrapper[4758]: I1203 18:14:10.114223 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:14:10 crc kubenswrapper[4758]: E1203 18:14:10.114914 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:14:10 crc kubenswrapper[4758]: I1203 18:14:10.534677 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hbx7w" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="registry-server" containerID="cri-o://6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3" gracePeriod=2 Dec 03 18:14:11 crc kubenswrapper[4758]: I1203 18:14:11.990425 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.168306 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-catalog-content\") pod \"d38dca61-3c59-47c9-9605-3cf815add7bf\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.168600 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-utilities\") pod \"d38dca61-3c59-47c9-9605-3cf815add7bf\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.169111 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw4n8\" (UniqueName: \"kubernetes.io/projected/d38dca61-3c59-47c9-9605-3cf815add7bf-kube-api-access-xw4n8\") pod \"d38dca61-3c59-47c9-9605-3cf815add7bf\" (UID: \"d38dca61-3c59-47c9-9605-3cf815add7bf\") " Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.169621 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-utilities" (OuterVolumeSpecName: "utilities") pod "d38dca61-3c59-47c9-9605-3cf815add7bf" (UID: "d38dca61-3c59-47c9-9605-3cf815add7bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.169819 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.174892 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38dca61-3c59-47c9-9605-3cf815add7bf-kube-api-access-xw4n8" (OuterVolumeSpecName: "kube-api-access-xw4n8") pod "d38dca61-3c59-47c9-9605-3cf815add7bf" (UID: "d38dca61-3c59-47c9-9605-3cf815add7bf"). InnerVolumeSpecName "kube-api-access-xw4n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.271204 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw4n8\" (UniqueName: \"kubernetes.io/projected/d38dca61-3c59-47c9-9605-3cf815add7bf-kube-api-access-xw4n8\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.298907 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d38dca61-3c59-47c9-9605-3cf815add7bf" (UID: "d38dca61-3c59-47c9-9605-3cf815add7bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.372485 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38dca61-3c59-47c9-9605-3cf815add7bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.550551 4758 generic.go:334] "Generic (PLEG): container finished" podID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerID="6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3" exitCode=0 Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.550590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerDied","Data":"6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3"} Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.550611 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbx7w" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.550628 4758 scope.go:117] "RemoveContainer" containerID="6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.550614 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbx7w" event={"ID":"d38dca61-3c59-47c9-9605-3cf815add7bf","Type":"ContainerDied","Data":"5d87fe81bd79b8b00ae4b2f623a23ed6d0ad12a97dd335bd98aa40c0fe46303b"} Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.589927 4758 scope.go:117] "RemoveContainer" containerID="6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.590119 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbx7w"] Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.598390 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hbx7w"] Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.616635 4758 scope.go:117] "RemoveContainer" containerID="7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.642846 4758 scope.go:117] "RemoveContainer" containerID="6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3" Dec 03 18:14:12 crc kubenswrapper[4758]: E1203 18:14:12.643289 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3\": container with ID starting with 6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3 not found: ID does not exist" containerID="6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.643321 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3"} err="failed to get container status \"6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3\": rpc error: code = NotFound desc = could not find container \"6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3\": container with ID starting with 6bf3888a123485928d88776a17f30db7a96955d0adab40ebb48e7301b267a7c3 not found: ID does not exist" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.643343 4758 scope.go:117] "RemoveContainer" containerID="6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765" Dec 03 18:14:12 crc kubenswrapper[4758]: E1203 18:14:12.643629 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765\": container with ID starting with 6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765 not found: ID does not exist" containerID="6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.643675 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765"} err="failed to get container status \"6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765\": rpc error: code = NotFound desc = could not find container \"6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765\": container with ID starting with 6d9dd3c258f8a0940c4032f008cb6afaf38cceda970cb54d6887aa77fa724765 not found: ID does not exist" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.643760 4758 scope.go:117] "RemoveContainer" containerID="7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2" Dec 03 18:14:12 crc kubenswrapper[4758]: E1203 18:14:12.644018 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2\": container with ID starting with 7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2 not found: ID does not exist" containerID="7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2" Dec 03 18:14:12 crc kubenswrapper[4758]: I1203 18:14:12.644039 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2"} err="failed to get container status \"7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2\": rpc error: code = NotFound desc = could not find container \"7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2\": container with ID starting with 7ed60243d7836c115d4dc1bf31de7f13dc8aa201612399250ecaf7f66df88cb2 not found: ID does not exist" Dec 03 18:14:13 crc kubenswrapper[4758]: I1203 18:14:13.124762 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" path="/var/lib/kubelet/pods/d38dca61-3c59-47c9-9605-3cf815add7bf/volumes" Dec 03 18:14:25 crc kubenswrapper[4758]: I1203 18:14:25.118304 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:14:25 crc kubenswrapper[4758]: I1203 18:14:25.650043 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"9807157ac08bd499c3c2865cca4e64a21569ebe37d2458c786adcfa7ed1dcacf"} Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.145589 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz"] Dec 03 18:15:00 crc kubenswrapper[4758]: E1203 18:15:00.146587 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="extract-utilities" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.146620 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="extract-utilities" Dec 03 18:15:00 crc kubenswrapper[4758]: E1203 18:15:00.146646 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="registry-server" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.146653 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="registry-server" Dec 03 18:15:00 crc kubenswrapper[4758]: E1203 18:15:00.146666 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="extract-content" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.146674 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="extract-content" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.146879 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38dca61-3c59-47c9-9605-3cf815add7bf" containerName="registry-server" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.147581 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.150217 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.150478 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.155438 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz"] Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.192902 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52zlj\" (UniqueName: \"kubernetes.io/projected/de23699d-45ea-4220-8b62-6214fd0f33b5-kube-api-access-52zlj\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.192969 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de23699d-45ea-4220-8b62-6214fd0f33b5-config-volume\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.193082 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de23699d-45ea-4220-8b62-6214fd0f33b5-secret-volume\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.294921 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52zlj\" (UniqueName: \"kubernetes.io/projected/de23699d-45ea-4220-8b62-6214fd0f33b5-kube-api-access-52zlj\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.295045 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de23699d-45ea-4220-8b62-6214fd0f33b5-config-volume\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.296214 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de23699d-45ea-4220-8b62-6214fd0f33b5-config-volume\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.296401 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de23699d-45ea-4220-8b62-6214fd0f33b5-secret-volume\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.310076 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de23699d-45ea-4220-8b62-6214fd0f33b5-secret-volume\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.314182 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52zlj\" (UniqueName: \"kubernetes.io/projected/de23699d-45ea-4220-8b62-6214fd0f33b5-kube-api-access-52zlj\") pod \"collect-profiles-29413095-6w4pz\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.469000 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:00 crc kubenswrapper[4758]: I1203 18:15:00.875019 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz"] Dec 03 18:15:00 crc kubenswrapper[4758]: W1203 18:15:00.938732 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde23699d_45ea_4220_8b62_6214fd0f33b5.slice/crio-78246642dd7487c42a148196a89e78cdd108508019b77cd49ccad8dc5725ecd8 WatchSource:0}: Error finding container 78246642dd7487c42a148196a89e78cdd108508019b77cd49ccad8dc5725ecd8: Status 404 returned error can't find the container with id 78246642dd7487c42a148196a89e78cdd108508019b77cd49ccad8dc5725ecd8 Dec 03 18:15:01 crc kubenswrapper[4758]: I1203 18:15:01.933397 4758 generic.go:334] "Generic (PLEG): container finished" podID="de23699d-45ea-4220-8b62-6214fd0f33b5" containerID="0d07c9f24299f0fe2ec54708ac70d5cd823422490cb1e05af0c98d7e218cd143" exitCode=0 Dec 03 18:15:01 crc kubenswrapper[4758]: I1203 18:15:01.933500 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" event={"ID":"de23699d-45ea-4220-8b62-6214fd0f33b5","Type":"ContainerDied","Data":"0d07c9f24299f0fe2ec54708ac70d5cd823422490cb1e05af0c98d7e218cd143"} Dec 03 18:15:01 crc kubenswrapper[4758]: I1203 18:15:01.934771 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" event={"ID":"de23699d-45ea-4220-8b62-6214fd0f33b5","Type":"ContainerStarted","Data":"78246642dd7487c42a148196a89e78cdd108508019b77cd49ccad8dc5725ecd8"} Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.193491 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.248770 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de23699d-45ea-4220-8b62-6214fd0f33b5-secret-volume\") pod \"de23699d-45ea-4220-8b62-6214fd0f33b5\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.248903 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52zlj\" (UniqueName: \"kubernetes.io/projected/de23699d-45ea-4220-8b62-6214fd0f33b5-kube-api-access-52zlj\") pod \"de23699d-45ea-4220-8b62-6214fd0f33b5\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.249010 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de23699d-45ea-4220-8b62-6214fd0f33b5-config-volume\") pod \"de23699d-45ea-4220-8b62-6214fd0f33b5\" (UID: \"de23699d-45ea-4220-8b62-6214fd0f33b5\") " Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.249608 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de23699d-45ea-4220-8b62-6214fd0f33b5-config-volume" (OuterVolumeSpecName: "config-volume") pod "de23699d-45ea-4220-8b62-6214fd0f33b5" (UID: "de23699d-45ea-4220-8b62-6214fd0f33b5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.256478 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de23699d-45ea-4220-8b62-6214fd0f33b5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "de23699d-45ea-4220-8b62-6214fd0f33b5" (UID: "de23699d-45ea-4220-8b62-6214fd0f33b5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.256715 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de23699d-45ea-4220-8b62-6214fd0f33b5-kube-api-access-52zlj" (OuterVolumeSpecName: "kube-api-access-52zlj") pod "de23699d-45ea-4220-8b62-6214fd0f33b5" (UID: "de23699d-45ea-4220-8b62-6214fd0f33b5"). InnerVolumeSpecName "kube-api-access-52zlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.350891 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de23699d-45ea-4220-8b62-6214fd0f33b5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.351020 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52zlj\" (UniqueName: \"kubernetes.io/projected/de23699d-45ea-4220-8b62-6214fd0f33b5-kube-api-access-52zlj\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.351038 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de23699d-45ea-4220-8b62-6214fd0f33b5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.950024 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" event={"ID":"de23699d-45ea-4220-8b62-6214fd0f33b5","Type":"ContainerDied","Data":"78246642dd7487c42a148196a89e78cdd108508019b77cd49ccad8dc5725ecd8"} Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.950383 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78246642dd7487c42a148196a89e78cdd108508019b77cd49ccad8dc5725ecd8" Dec 03 18:15:03 crc kubenswrapper[4758]: I1203 18:15:03.950091 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz" Dec 03 18:15:04 crc kubenswrapper[4758]: I1203 18:15:04.262453 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6"] Dec 03 18:15:04 crc kubenswrapper[4758]: I1203 18:15:04.267825 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-dsbm6"] Dec 03 18:15:05 crc kubenswrapper[4758]: I1203 18:15:05.133041 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee99b637-61db-4e76-8ef8-05da30ba956b" path="/var/lib/kubelet/pods/ee99b637-61db-4e76-8ef8-05da30ba956b/volumes" Dec 03 18:15:10 crc kubenswrapper[4758]: I1203 18:15:10.402562 4758 scope.go:117] "RemoveContainer" containerID="7fbb00cf742a12cdf4376c50ef72949a4d018f1ef4d49fa3ce11599f1a8f7c66" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.302397 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ppjsr"] Dec 03 18:15:22 crc kubenswrapper[4758]: E1203 18:15:22.303304 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de23699d-45ea-4220-8b62-6214fd0f33b5" containerName="collect-profiles" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.303320 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="de23699d-45ea-4220-8b62-6214fd0f33b5" containerName="collect-profiles" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.303508 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="de23699d-45ea-4220-8b62-6214fd0f33b5" containerName="collect-profiles" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.304735 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.313194 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppjsr"] Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.364806 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-utilities\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.364889 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shqwz\" (UniqueName: \"kubernetes.io/projected/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-kube-api-access-shqwz\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.364938 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-catalog-content\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.466853 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-utilities\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.466994 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shqwz\" (UniqueName: \"kubernetes.io/projected/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-kube-api-access-shqwz\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.467065 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-catalog-content\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.467465 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-utilities\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.467497 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-catalog-content\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.495882 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shqwz\" (UniqueName: \"kubernetes.io/projected/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-kube-api-access-shqwz\") pod \"community-operators-ppjsr\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:22 crc kubenswrapper[4758]: I1203 18:15:22.628109 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:23 crc kubenswrapper[4758]: I1203 18:15:23.122892 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppjsr"] Dec 03 18:15:23 crc kubenswrapper[4758]: I1203 18:15:23.160119 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppjsr" event={"ID":"a4a9e5c8-7003-41a3-8882-cded31c3b8f4","Type":"ContainerStarted","Data":"88354873acc766e3a9b803b76256c2383132125957c9ebc27ef5dae0e2b0a3c3"} Dec 03 18:15:24 crc kubenswrapper[4758]: I1203 18:15:24.167402 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerID="62a0561cd808ecd30a2d408643ac51f49c4bafd23ff33631c7bb9eb46d59a88d" exitCode=0 Dec 03 18:15:24 crc kubenswrapper[4758]: I1203 18:15:24.167505 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppjsr" event={"ID":"a4a9e5c8-7003-41a3-8882-cded31c3b8f4","Type":"ContainerDied","Data":"62a0561cd808ecd30a2d408643ac51f49c4bafd23ff33631c7bb9eb46d59a88d"} Dec 03 18:15:26 crc kubenswrapper[4758]: I1203 18:15:26.185846 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerID="83692357a0951245097d8fb23ceb4f1bffbe7dd82e1bf0893397cbe563127974" exitCode=0 Dec 03 18:15:26 crc kubenswrapper[4758]: I1203 18:15:26.185957 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppjsr" event={"ID":"a4a9e5c8-7003-41a3-8882-cded31c3b8f4","Type":"ContainerDied","Data":"83692357a0951245097d8fb23ceb4f1bffbe7dd82e1bf0893397cbe563127974"} Dec 03 18:15:27 crc kubenswrapper[4758]: I1203 18:15:27.200640 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppjsr" event={"ID":"a4a9e5c8-7003-41a3-8882-cded31c3b8f4","Type":"ContainerStarted","Data":"2d590e2baf62564bea530b32c430afb8853df4d06583bd1e3ddbebcc7150dd0b"} Dec 03 18:15:27 crc kubenswrapper[4758]: I1203 18:15:27.228844 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ppjsr" podStartSLOduration=2.796179335 podStartE2EDuration="5.228824299s" podCreationTimestamp="2025-12-03 18:15:22 +0000 UTC" firstStartedPulling="2025-12-03 18:15:24.168861455 +0000 UTC m=+4779.370238316" lastFinishedPulling="2025-12-03 18:15:26.601506419 +0000 UTC m=+4781.802883280" observedRunningTime="2025-12-03 18:15:27.221287937 +0000 UTC m=+4782.422664818" watchObservedRunningTime="2025-12-03 18:15:27.228824299 +0000 UTC m=+4782.430201160" Dec 03 18:15:32 crc kubenswrapper[4758]: I1203 18:15:32.629195 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:32 crc kubenswrapper[4758]: I1203 18:15:32.630242 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:32 crc kubenswrapper[4758]: I1203 18:15:32.695205 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:33 crc kubenswrapper[4758]: I1203 18:15:33.300202 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:33 crc kubenswrapper[4758]: I1203 18:15:33.347800 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppjsr"] Dec 03 18:15:35 crc kubenswrapper[4758]: I1203 18:15:35.261257 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ppjsr" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="registry-server" containerID="cri-o://2d590e2baf62564bea530b32c430afb8853df4d06583bd1e3ddbebcc7150dd0b" gracePeriod=2 Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.267479 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerID="2d590e2baf62564bea530b32c430afb8853df4d06583bd1e3ddbebcc7150dd0b" exitCode=0 Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.267517 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppjsr" event={"ID":"a4a9e5c8-7003-41a3-8882-cded31c3b8f4","Type":"ContainerDied","Data":"2d590e2baf62564bea530b32c430afb8853df4d06583bd1e3ddbebcc7150dd0b"} Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.793021 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.971808 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-utilities\") pod \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.971893 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shqwz\" (UniqueName: \"kubernetes.io/projected/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-kube-api-access-shqwz\") pod \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.972022 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-catalog-content\") pod \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\" (UID: \"a4a9e5c8-7003-41a3-8882-cded31c3b8f4\") " Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.973534 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-utilities" (OuterVolumeSpecName: "utilities") pod "a4a9e5c8-7003-41a3-8882-cded31c3b8f4" (UID: "a4a9e5c8-7003-41a3-8882-cded31c3b8f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:15:36 crc kubenswrapper[4758]: I1203 18:15:36.990846 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-kube-api-access-shqwz" (OuterVolumeSpecName: "kube-api-access-shqwz") pod "a4a9e5c8-7003-41a3-8882-cded31c3b8f4" (UID: "a4a9e5c8-7003-41a3-8882-cded31c3b8f4"). InnerVolumeSpecName "kube-api-access-shqwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.047407 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4a9e5c8-7003-41a3-8882-cded31c3b8f4" (UID: "a4a9e5c8-7003-41a3-8882-cded31c3b8f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.073340 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.073389 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shqwz\" (UniqueName: \"kubernetes.io/projected/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-kube-api-access-shqwz\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.073408 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a9e5c8-7003-41a3-8882-cded31c3b8f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.278426 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppjsr" event={"ID":"a4a9e5c8-7003-41a3-8882-cded31c3b8f4","Type":"ContainerDied","Data":"88354873acc766e3a9b803b76256c2383132125957c9ebc27ef5dae0e2b0a3c3"} Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.278666 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppjsr" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.278801 4758 scope.go:117] "RemoveContainer" containerID="2d590e2baf62564bea530b32c430afb8853df4d06583bd1e3ddbebcc7150dd0b" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.306419 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppjsr"] Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.310655 4758 scope.go:117] "RemoveContainer" containerID="83692357a0951245097d8fb23ceb4f1bffbe7dd82e1bf0893397cbe563127974" Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.314566 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ppjsr"] Dec 03 18:15:37 crc kubenswrapper[4758]: I1203 18:15:37.349760 4758 scope.go:117] "RemoveContainer" containerID="62a0561cd808ecd30a2d408643ac51f49c4bafd23ff33631c7bb9eb46d59a88d" Dec 03 18:15:39 crc kubenswrapper[4758]: I1203 18:15:39.129663 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" path="/var/lib/kubelet/pods/a4a9e5c8-7003-41a3-8882-cded31c3b8f4/volumes" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.348757 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-mhpzq"] Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.373514 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-mhpzq"] Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.495897 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-xgvh9"] Dec 03 18:15:41 crc kubenswrapper[4758]: E1203 18:15:41.498993 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="registry-server" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.499032 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="registry-server" Dec 03 18:15:41 crc kubenswrapper[4758]: E1203 18:15:41.499053 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="extract-content" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.499061 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="extract-content" Dec 03 18:15:41 crc kubenswrapper[4758]: E1203 18:15:41.499081 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="extract-utilities" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.499089 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="extract-utilities" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.499341 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a9e5c8-7003-41a3-8882-cded31c3b8f4" containerName="registry-server" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.500056 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.500659 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-xgvh9"] Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.502610 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.502813 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.503101 4758 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-ggqk7" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.504563 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.655708 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsc4d\" (UniqueName: \"kubernetes.io/projected/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-kube-api-access-wsc4d\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.655751 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-node-mnt\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.655833 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-crc-storage\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.757762 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsc4d\" (UniqueName: \"kubernetes.io/projected/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-kube-api-access-wsc4d\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.757832 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-node-mnt\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.758282 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-node-mnt\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.758489 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-crc-storage\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.759812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-crc-storage\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.780574 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsc4d\" (UniqueName: \"kubernetes.io/projected/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-kube-api-access-wsc4d\") pod \"crc-storage-crc-xgvh9\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:41 crc kubenswrapper[4758]: I1203 18:15:41.819868 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:42 crc kubenswrapper[4758]: I1203 18:15:42.334151 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-xgvh9"] Dec 03 18:15:43 crc kubenswrapper[4758]: I1203 18:15:43.134871 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85230d29-9148-4067-a103-2482c48013fb" path="/var/lib/kubelet/pods/85230d29-9148-4067-a103-2482c48013fb/volumes" Dec 03 18:15:43 crc kubenswrapper[4758]: I1203 18:15:43.336567 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xgvh9" event={"ID":"46f6bce5-09c7-4086-8b4d-b0b9d10d557f","Type":"ContainerStarted","Data":"02f2c53787ea82adbb557c54ba14b4e72c526fe6a9d1dd1ad2cff6b3f9b5e37d"} Dec 03 18:15:43 crc kubenswrapper[4758]: I1203 18:15:43.336632 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xgvh9" event={"ID":"46f6bce5-09c7-4086-8b4d-b0b9d10d557f","Type":"ContainerStarted","Data":"cbeb5638f0301956dc0f5022e10639941618186d3a11645cb9ed35770c1e84d6"} Dec 03 18:15:43 crc kubenswrapper[4758]: I1203 18:15:43.368227 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-xgvh9" podStartSLOduration=1.820000093 podStartE2EDuration="2.368199953s" podCreationTimestamp="2025-12-03 18:15:41 +0000 UTC" firstStartedPulling="2025-12-03 18:15:42.33958912 +0000 UTC m=+4797.540966031" lastFinishedPulling="2025-12-03 18:15:42.88778903 +0000 UTC m=+4798.089165891" observedRunningTime="2025-12-03 18:15:43.357974889 +0000 UTC m=+4798.559351770" watchObservedRunningTime="2025-12-03 18:15:43.368199953 +0000 UTC m=+4798.569576844" Dec 03 18:15:44 crc kubenswrapper[4758]: I1203 18:15:44.348110 4758 generic.go:334] "Generic (PLEG): container finished" podID="46f6bce5-09c7-4086-8b4d-b0b9d10d557f" containerID="02f2c53787ea82adbb557c54ba14b4e72c526fe6a9d1dd1ad2cff6b3f9b5e37d" exitCode=0 Dec 03 18:15:44 crc kubenswrapper[4758]: I1203 18:15:44.348156 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xgvh9" event={"ID":"46f6bce5-09c7-4086-8b4d-b0b9d10d557f","Type":"ContainerDied","Data":"02f2c53787ea82adbb557c54ba14b4e72c526fe6a9d1dd1ad2cff6b3f9b5e37d"} Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.728639 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.822435 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-crc-storage\") pod \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.822877 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsc4d\" (UniqueName: \"kubernetes.io/projected/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-kube-api-access-wsc4d\") pod \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.823102 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-node-mnt\") pod \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\" (UID: \"46f6bce5-09c7-4086-8b4d-b0b9d10d557f\") " Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.823222 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "46f6bce5-09c7-4086-8b4d-b0b9d10d557f" (UID: "46f6bce5-09c7-4086-8b4d-b0b9d10d557f"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.824008 4758 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.828252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-kube-api-access-wsc4d" (OuterVolumeSpecName: "kube-api-access-wsc4d") pod "46f6bce5-09c7-4086-8b4d-b0b9d10d557f" (UID: "46f6bce5-09c7-4086-8b4d-b0b9d10d557f"). InnerVolumeSpecName "kube-api-access-wsc4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.840779 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "46f6bce5-09c7-4086-8b4d-b0b9d10d557f" (UID: "46f6bce5-09c7-4086-8b4d-b0b9d10d557f"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.926419 4758 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:45 crc kubenswrapper[4758]: I1203 18:15:45.926496 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsc4d\" (UniqueName: \"kubernetes.io/projected/46f6bce5-09c7-4086-8b4d-b0b9d10d557f-kube-api-access-wsc4d\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:46 crc kubenswrapper[4758]: I1203 18:15:46.370399 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xgvh9" event={"ID":"46f6bce5-09c7-4086-8b4d-b0b9d10d557f","Type":"ContainerDied","Data":"cbeb5638f0301956dc0f5022e10639941618186d3a11645cb9ed35770c1e84d6"} Dec 03 18:15:46 crc kubenswrapper[4758]: I1203 18:15:46.370774 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbeb5638f0301956dc0f5022e10639941618186d3a11645cb9ed35770c1e84d6" Dec 03 18:15:46 crc kubenswrapper[4758]: I1203 18:15:46.370537 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xgvh9" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.702372 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-xgvh9"] Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.709311 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-xgvh9"] Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.843000 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-mnh56"] Dec 03 18:15:47 crc kubenswrapper[4758]: E1203 18:15:47.844042 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f6bce5-09c7-4086-8b4d-b0b9d10d557f" containerName="storage" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.844075 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f6bce5-09c7-4086-8b4d-b0b9d10d557f" containerName="storage" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.844258 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f6bce5-09c7-4086-8b4d-b0b9d10d557f" containerName="storage" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.844877 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.847767 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.847839 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.847881 4758 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-ggqk7" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.853833 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.856940 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-mnh56"] Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.957339 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhfs2\" (UniqueName: \"kubernetes.io/projected/6f086617-643a-4694-ace9-5b2248bcf49b-kube-api-access-jhfs2\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.957440 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f086617-643a-4694-ace9-5b2248bcf49b-crc-storage\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:47 crc kubenswrapper[4758]: I1203 18:15:47.957531 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f086617-643a-4694-ace9-5b2248bcf49b-node-mnt\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.058853 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f086617-643a-4694-ace9-5b2248bcf49b-node-mnt\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.058966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhfs2\" (UniqueName: \"kubernetes.io/projected/6f086617-643a-4694-ace9-5b2248bcf49b-kube-api-access-jhfs2\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.058994 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f086617-643a-4694-ace9-5b2248bcf49b-crc-storage\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.059144 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f086617-643a-4694-ace9-5b2248bcf49b-node-mnt\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.059644 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f086617-643a-4694-ace9-5b2248bcf49b-crc-storage\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.086536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhfs2\" (UniqueName: \"kubernetes.io/projected/6f086617-643a-4694-ace9-5b2248bcf49b-kube-api-access-jhfs2\") pod \"crc-storage-crc-mnh56\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.168084 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:48 crc kubenswrapper[4758]: I1203 18:15:48.423797 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-mnh56"] Dec 03 18:15:49 crc kubenswrapper[4758]: I1203 18:15:49.122060 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f6bce5-09c7-4086-8b4d-b0b9d10d557f" path="/var/lib/kubelet/pods/46f6bce5-09c7-4086-8b4d-b0b9d10d557f/volumes" Dec 03 18:15:49 crc kubenswrapper[4758]: I1203 18:15:49.393336 4758 generic.go:334] "Generic (PLEG): container finished" podID="6f086617-643a-4694-ace9-5b2248bcf49b" containerID="88ef6c1add6ad27bafbe4f13a42fc69c6b30b984af7cb82e165686aa7df532c4" exitCode=0 Dec 03 18:15:49 crc kubenswrapper[4758]: I1203 18:15:49.393489 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-mnh56" event={"ID":"6f086617-643a-4694-ace9-5b2248bcf49b","Type":"ContainerDied","Data":"88ef6c1add6ad27bafbe4f13a42fc69c6b30b984af7cb82e165686aa7df532c4"} Dec 03 18:15:49 crc kubenswrapper[4758]: I1203 18:15:49.393519 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-mnh56" event={"ID":"6f086617-643a-4694-ace9-5b2248bcf49b","Type":"ContainerStarted","Data":"6b198460ab3be273f9f0c6ea760786dbbe61e070b0a561ab10e765f5e17d3ee3"} Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.742855 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.900088 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f086617-643a-4694-ace9-5b2248bcf49b-crc-storage\") pod \"6f086617-643a-4694-ace9-5b2248bcf49b\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.900184 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f086617-643a-4694-ace9-5b2248bcf49b-node-mnt\") pod \"6f086617-643a-4694-ace9-5b2248bcf49b\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.900252 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhfs2\" (UniqueName: \"kubernetes.io/projected/6f086617-643a-4694-ace9-5b2248bcf49b-kube-api-access-jhfs2\") pod \"6f086617-643a-4694-ace9-5b2248bcf49b\" (UID: \"6f086617-643a-4694-ace9-5b2248bcf49b\") " Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.900340 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f086617-643a-4694-ace9-5b2248bcf49b-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "6f086617-643a-4694-ace9-5b2248bcf49b" (UID: "6f086617-643a-4694-ace9-5b2248bcf49b"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.900611 4758 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f086617-643a-4694-ace9-5b2248bcf49b-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.906526 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f086617-643a-4694-ace9-5b2248bcf49b-kube-api-access-jhfs2" (OuterVolumeSpecName: "kube-api-access-jhfs2") pod "6f086617-643a-4694-ace9-5b2248bcf49b" (UID: "6f086617-643a-4694-ace9-5b2248bcf49b"). InnerVolumeSpecName "kube-api-access-jhfs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:15:50 crc kubenswrapper[4758]: I1203 18:15:50.930222 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f086617-643a-4694-ace9-5b2248bcf49b-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "6f086617-643a-4694-ace9-5b2248bcf49b" (UID: "6f086617-643a-4694-ace9-5b2248bcf49b"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:15:51 crc kubenswrapper[4758]: I1203 18:15:51.002719 4758 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f086617-643a-4694-ace9-5b2248bcf49b-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:51 crc kubenswrapper[4758]: I1203 18:15:51.002771 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhfs2\" (UniqueName: \"kubernetes.io/projected/6f086617-643a-4694-ace9-5b2248bcf49b-kube-api-access-jhfs2\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:51 crc kubenswrapper[4758]: I1203 18:15:51.420920 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-mnh56" event={"ID":"6f086617-643a-4694-ace9-5b2248bcf49b","Type":"ContainerDied","Data":"6b198460ab3be273f9f0c6ea760786dbbe61e070b0a561ab10e765f5e17d3ee3"} Dec 03 18:15:51 crc kubenswrapper[4758]: I1203 18:15:51.420994 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b198460ab3be273f9f0c6ea760786dbbe61e070b0a561ab10e765f5e17d3ee3" Dec 03 18:15:51 crc kubenswrapper[4758]: I1203 18:15:51.421101 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-mnh56" Dec 03 18:16:10 crc kubenswrapper[4758]: I1203 18:16:10.500622 4758 scope.go:117] "RemoveContainer" containerID="9b4c1858e6eb9e695919edccf921e60150fbafa35f743e2d599dee47e401775a" Dec 03 18:16:41 crc kubenswrapper[4758]: I1203 18:16:41.394759 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:16:41 crc kubenswrapper[4758]: I1203 18:16:41.395223 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:17:11 crc kubenswrapper[4758]: I1203 18:17:11.394351 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:17:11 crc kubenswrapper[4758]: I1203 18:17:11.394962 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:17:41 crc kubenswrapper[4758]: I1203 18:17:41.394858 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:17:41 crc kubenswrapper[4758]: I1203 18:17:41.395548 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:17:41 crc kubenswrapper[4758]: I1203 18:17:41.395615 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:17:41 crc kubenswrapper[4758]: I1203 18:17:41.396586 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9807157ac08bd499c3c2865cca4e64a21569ebe37d2458c786adcfa7ed1dcacf"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:17:41 crc kubenswrapper[4758]: I1203 18:17:41.396674 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://9807157ac08bd499c3c2865cca4e64a21569ebe37d2458c786adcfa7ed1dcacf" gracePeriod=600 Dec 03 18:17:42 crc kubenswrapper[4758]: I1203 18:17:42.289971 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="9807157ac08bd499c3c2865cca4e64a21569ebe37d2458c786adcfa7ed1dcacf" exitCode=0 Dec 03 18:17:42 crc kubenswrapper[4758]: I1203 18:17:42.290043 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"9807157ac08bd499c3c2865cca4e64a21569ebe37d2458c786adcfa7ed1dcacf"} Dec 03 18:17:42 crc kubenswrapper[4758]: I1203 18:17:42.290580 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176"} Dec 03 18:17:42 crc kubenswrapper[4758]: I1203 18:17:42.290602 4758 scope.go:117] "RemoveContainer" containerID="d297f6f56c5e207ec1d509be69f195fbe2efed090aab3bd236d0bf9c510b2949" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.768452 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-5nlxk"] Dec 03 18:18:57 crc kubenswrapper[4758]: E1203 18:18:57.769260 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f086617-643a-4694-ace9-5b2248bcf49b" containerName="storage" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.769273 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f086617-643a-4694-ace9-5b2248bcf49b" containerName="storage" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.769408 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f086617-643a-4694-ace9-5b2248bcf49b" containerName="storage" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.770117 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.773715 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.774012 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.776195 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-lxdf2" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.776418 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.776564 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.789702 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-5nlxk"] Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.920271 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.920561 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvh77\" (UniqueName: \"kubernetes.io/projected/9a35df35-4f43-4720-b12a-b393f9f56c64-kube-api-access-fvh77\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.920593 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-config\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.954351 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-t7h6f"] Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.961798 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:57 crc kubenswrapper[4758]: I1203 18:18:57.964655 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-t7h6f"] Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.022444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvh77\" (UniqueName: \"kubernetes.io/projected/9a35df35-4f43-4720-b12a-b393f9f56c64-kube-api-access-fvh77\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.022501 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-config\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.022605 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.023558 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.023577 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-config\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.058829 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvh77\" (UniqueName: \"kubernetes.io/projected/9a35df35-4f43-4720-b12a-b393f9f56c64-kube-api-access-fvh77\") pod \"dnsmasq-dns-5d7b5456f5-5nlxk\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.091184 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.123485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.123529 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-config\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.123589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfnhc\" (UniqueName: \"kubernetes.io/projected/073714d5-1c99-44bf-8bf2-d14dc3a08e32-kube-api-access-bfnhc\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.225207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.225453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-config\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.225515 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfnhc\" (UniqueName: \"kubernetes.io/projected/073714d5-1c99-44bf-8bf2-d14dc3a08e32-kube-api-access-bfnhc\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.226295 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.226455 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-config\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.250664 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfnhc\" (UniqueName: \"kubernetes.io/projected/073714d5-1c99-44bf-8bf2-d14dc3a08e32-kube-api-access-bfnhc\") pod \"dnsmasq-dns-98ddfc8f-t7h6f\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.296958 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.539490 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-5nlxk"] Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.698067 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-t7h6f"] Dec 03 18:18:58 crc kubenswrapper[4758]: W1203 18:18:58.699078 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod073714d5_1c99_44bf_8bf2_d14dc3a08e32.slice/crio-3f1d509dc49917f6cf3d47d600402fc5798c31d035dfb82ef8200a75d0f20a90 WatchSource:0}: Error finding container 3f1d509dc49917f6cf3d47d600402fc5798c31d035dfb82ef8200a75d0f20a90: Status 404 returned error can't find the container with id 3f1d509dc49917f6cf3d47d600402fc5798c31d035dfb82ef8200a75d0f20a90 Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.866041 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.867372 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.869570 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.869699 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.869719 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.871510 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.871762 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sb5gc" Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.884322 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.960084 4758 generic.go:334] "Generic (PLEG): container finished" podID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerID="dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc" exitCode=0 Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.960187 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" event={"ID":"9a35df35-4f43-4720-b12a-b393f9f56c64","Type":"ContainerDied","Data":"dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc"} Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.960238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" event={"ID":"9a35df35-4f43-4720-b12a-b393f9f56c64","Type":"ContainerStarted","Data":"ac16cc7b9d3195aaa32b4fe157573efebc8c87fa472839d01fd62d73ce423bdb"} Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.961916 4758 generic.go:334] "Generic (PLEG): container finished" podID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerID="fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa" exitCode=0 Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.962085 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" event={"ID":"073714d5-1c99-44bf-8bf2-d14dc3a08e32","Type":"ContainerDied","Data":"fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa"} Dec 03 18:18:58 crc kubenswrapper[4758]: I1203 18:18:58.962169 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" event={"ID":"073714d5-1c99-44bf-8bf2-d14dc3a08e32","Type":"ContainerStarted","Data":"3f1d509dc49917f6cf3d47d600402fc5798c31d035dfb82ef8200a75d0f20a90"} Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037088 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037165 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037264 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/386626ba-ab82-4063-880e-96518ca1426a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037335 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037374 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/386626ba-ab82-4063-880e-96518ca1426a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037406 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45fdn\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-kube-api-access-45fdn\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037740 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.037828 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.139761 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.140130 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/386626ba-ab82-4063-880e-96518ca1426a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.140156 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45fdn\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-kube-api-access-45fdn\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.140233 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.140290 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.141146 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.142894 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.142974 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.143005 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.143074 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/386626ba-ab82-4063-880e-96518ca1426a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.143957 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.145109 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.145115 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.145271 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/386626ba-ab82-4063-880e-96518ca1426a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.145288 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.147691 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.149162 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.151406 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/386626ba-ab82-4063-880e-96518ca1426a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.152257 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.152451 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.152571 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.152616 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.152643 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7xm5w" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.156322 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.156357 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4213de258debccea79f1d56b24d3aacf86159bcd6ca8adb0d34458dcf4f780d4/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.165436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45fdn\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-kube-api-access-45fdn\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.167239 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.221876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.346870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.346962 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347064 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347213 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347250 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/165e9da9-68ad-45ab-b162-9e70cea488d2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347275 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/165e9da9-68ad-45ab-b162-9e70cea488d2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347511 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347571 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.347626 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dnwz\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-kube-api-access-6dnwz\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448608 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448656 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448711 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dnwz\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-kube-api-access-6dnwz\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448755 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448789 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448896 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448923 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/165e9da9-68ad-45ab-b162-9e70cea488d2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.448947 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/165e9da9-68ad-45ab-b162-9e70cea488d2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.449416 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.449613 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.450051 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.450732 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.450842 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.450883 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2613edf6caccdc4198d192ba5263c37a33bb7fb662d0fa7d899fb5d156d73312/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.453029 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.454185 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/165e9da9-68ad-45ab-b162-9e70cea488d2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.454276 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/165e9da9-68ad-45ab-b162-9e70cea488d2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.477820 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dnwz\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-kube-api-access-6dnwz\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.485496 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.494559 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.557451 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.805051 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.839284 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.969790 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"165e9da9-68ad-45ab-b162-9e70cea488d2","Type":"ContainerStarted","Data":"0da5450239decef359be067d28da09fb1aa702bb8a0e3490bfe073eef3bc02a6"} Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.972928 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" event={"ID":"073714d5-1c99-44bf-8bf2-d14dc3a08e32","Type":"ContainerStarted","Data":"1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180"} Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.973094 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.975229 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" event={"ID":"9a35df35-4f43-4720-b12a-b393f9f56c64","Type":"ContainerStarted","Data":"95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb"} Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.975639 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.976622 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"386626ba-ab82-4063-880e-96518ca1426a","Type":"ContainerStarted","Data":"4200e392eea3cb229dd00a930df7bcc461fa16f9d7edb195508c653575de031c"} Dec 03 18:18:59 crc kubenswrapper[4758]: I1203 18:18:59.992851 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" podStartSLOduration=2.992828368 podStartE2EDuration="2.992828368s" podCreationTimestamp="2025-12-03 18:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:18:59.988285106 +0000 UTC m=+4995.189661967" watchObservedRunningTime="2025-12-03 18:18:59.992828368 +0000 UTC m=+4995.194205249" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.010353 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" podStartSLOduration=3.01033497 podStartE2EDuration="3.01033497s" podCreationTimestamp="2025-12-03 18:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:00.004460332 +0000 UTC m=+4995.205837203" watchObservedRunningTime="2025-12-03 18:19:00.01033497 +0000 UTC m=+4995.211711831" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.353771 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.355093 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.359936 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.361665 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.363389 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.369142 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mmttd" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.369172 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.370347 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476620 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-config-data-default\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476672 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476724 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476744 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-kolla-config\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476760 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476792 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg4d9\" (UniqueName: \"kubernetes.io/projected/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-kube-api-access-xg4d9\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.476845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-config-data-default\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578173 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-kolla-config\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578273 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578348 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg4d9\" (UniqueName: \"kubernetes.io/projected/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-kube-api-access-xg4d9\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.578387 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.579025 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-config-data-default\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.579758 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.579924 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.580485 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-kolla-config\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.583168 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.583201 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf17c0e068d94e2322994d6216e839a9e8d098eb5b1898f1888f48561798dee4/globalmount\"" pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.583268 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.583395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.599700 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg4d9\" (UniqueName: \"kubernetes.io/projected/cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3-kube-api-access-xg4d9\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.614345 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fa49ae5-3160-4d03-9e40-4e425bd7523c\") pod \"openstack-galera-0\" (UID: \"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3\") " pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.683859 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.776001 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.777242 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.779399 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.779627 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-lfqdf" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.783235 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.883014 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/141e0710-b827-4f55-af39-16ef9ffc7efd-config-data\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.883084 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccm62\" (UniqueName: \"kubernetes.io/projected/141e0710-b827-4f55-af39-16ef9ffc7efd-kube-api-access-ccm62\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.883136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/141e0710-b827-4f55-af39-16ef9ffc7efd-kolla-config\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.984422 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/141e0710-b827-4f55-af39-16ef9ffc7efd-config-data\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.984798 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccm62\" (UniqueName: \"kubernetes.io/projected/141e0710-b827-4f55-af39-16ef9ffc7efd-kube-api-access-ccm62\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.984845 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/141e0710-b827-4f55-af39-16ef9ffc7efd-kolla-config\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.985342 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/141e0710-b827-4f55-af39-16ef9ffc7efd-config-data\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.985743 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"386626ba-ab82-4063-880e-96518ca1426a","Type":"ContainerStarted","Data":"dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56"} Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.985749 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/141e0710-b827-4f55-af39-16ef9ffc7efd-kolla-config\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:00 crc kubenswrapper[4758]: I1203 18:19:00.988602 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"165e9da9-68ad-45ab-b162-9e70cea488d2","Type":"ContainerStarted","Data":"5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e"} Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.010809 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccm62\" (UniqueName: \"kubernetes.io/projected/141e0710-b827-4f55-af39-16ef9ffc7efd-kube-api-access-ccm62\") pod \"memcached-0\" (UID: \"141e0710-b827-4f55-af39-16ef9ffc7efd\") " pod="openstack/memcached-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.231239 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.284936 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.699161 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 18:19:01 crc kubenswrapper[4758]: W1203 18:19:01.704873 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod141e0710_b827_4f55_af39_16ef9ffc7efd.slice/crio-119f24540ca38b266dd35643f8101fce7a44cd1eaacd442011796c7c76adcc0c WatchSource:0}: Error finding container 119f24540ca38b266dd35643f8101fce7a44cd1eaacd442011796c7c76adcc0c: Status 404 returned error can't find the container with id 119f24540ca38b266dd35643f8101fce7a44cd1eaacd442011796c7c76adcc0c Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.820463 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.822004 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.826258 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.829512 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-nc9dm" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.830974 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.831229 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.832813 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.899783 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.899887 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0bb80e50-176a-4fc0-890d-026294efeaaa-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.899923 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bb80e50-176a-4fc0-890d-026294efeaaa-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.899950 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.899977 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb80e50-176a-4fc0-890d-026294efeaaa-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.899993 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt49r\" (UniqueName: \"kubernetes.io/projected/0bb80e50-176a-4fc0-890d-026294efeaaa-kube-api-access-jt49r\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.900012 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.900032 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.997257 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"141e0710-b827-4f55-af39-16ef9ffc7efd","Type":"ContainerStarted","Data":"b030dee66c2263f284101236e780fc82258f786486ffbd21409234ef3cee808b"} Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.997296 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"141e0710-b827-4f55-af39-16ef9ffc7efd","Type":"ContainerStarted","Data":"119f24540ca38b266dd35643f8101fce7a44cd1eaacd442011796c7c76adcc0c"} Dec 03 18:19:01 crc kubenswrapper[4758]: I1203 18:19:01.997395 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.000188 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3","Type":"ContainerStarted","Data":"190eb954a0fa4291da6a9a386324313c5a9496bc0dd61c3922e52229e4eef3cf"} Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.000225 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3","Type":"ContainerStarted","Data":"b51b62dae065f6b519732d8242c3a49217b7f2b410cc6d572f72db0c0548952c"} Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.001927 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002108 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0bb80e50-176a-4fc0-890d-026294efeaaa-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002176 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bb80e50-176a-4fc0-890d-026294efeaaa-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002282 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb80e50-176a-4fc0-890d-026294efeaaa-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt49r\" (UniqueName: \"kubernetes.io/projected/0bb80e50-176a-4fc0-890d-026294efeaaa-kube-api-access-jt49r\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002357 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.002403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.003367 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0bb80e50-176a-4fc0-890d-026294efeaaa-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.004733 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.005932 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.006103 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0bb80e50-176a-4fc0-890d-026294efeaaa-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.006270 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.006309 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4b7e0883ea104e4821b6125aafc8445f654a69d40d9c94607d5b980bd657d6d0/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.009039 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb80e50-176a-4fc0-890d-026294efeaaa-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.010965 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bb80e50-176a-4fc0-890d-026294efeaaa-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.017451 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt49r\" (UniqueName: \"kubernetes.io/projected/0bb80e50-176a-4fc0-890d-026294efeaaa-kube-api-access-jt49r\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.029332 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.029311593 podStartE2EDuration="2.029311593s" podCreationTimestamp="2025-12-03 18:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:02.024309948 +0000 UTC m=+4997.225686809" watchObservedRunningTime="2025-12-03 18:19:02.029311593 +0000 UTC m=+4997.230688454" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.045953 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fe6b6b3-8629-47d9-9351-3b5298fb8a31\") pod \"openstack-cell1-galera-0\" (UID: \"0bb80e50-176a-4fc0-890d-026294efeaaa\") " pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.151069 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:02 crc kubenswrapper[4758]: I1203 18:19:02.554206 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 18:19:03 crc kubenswrapper[4758]: I1203 18:19:03.008568 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0bb80e50-176a-4fc0-890d-026294efeaaa","Type":"ContainerStarted","Data":"e1e823f10ed047575dc34eb793025dec42a500c114f36433cbf1c6e2c3ba06ff"} Dec 03 18:19:03 crc kubenswrapper[4758]: I1203 18:19:03.008958 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0bb80e50-176a-4fc0-890d-026294efeaaa","Type":"ContainerStarted","Data":"e81ee9e43d5b4ff5ad1863aec0fdd4f03e9e7d2c08e8415e97b8bcfaf699c656"} Dec 03 18:19:06 crc kubenswrapper[4758]: I1203 18:19:06.030309 4758 generic.go:334] "Generic (PLEG): container finished" podID="cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3" containerID="190eb954a0fa4291da6a9a386324313c5a9496bc0dd61c3922e52229e4eef3cf" exitCode=0 Dec 03 18:19:06 crc kubenswrapper[4758]: I1203 18:19:06.030398 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3","Type":"ContainerDied","Data":"190eb954a0fa4291da6a9a386324313c5a9496bc0dd61c3922e52229e4eef3cf"} Dec 03 18:19:06 crc kubenswrapper[4758]: I1203 18:19:06.286633 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 18:19:07 crc kubenswrapper[4758]: I1203 18:19:07.038552 4758 generic.go:334] "Generic (PLEG): container finished" podID="0bb80e50-176a-4fc0-890d-026294efeaaa" containerID="e1e823f10ed047575dc34eb793025dec42a500c114f36433cbf1c6e2c3ba06ff" exitCode=0 Dec 03 18:19:07 crc kubenswrapper[4758]: I1203 18:19:07.038863 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0bb80e50-176a-4fc0-890d-026294efeaaa","Type":"ContainerDied","Data":"e1e823f10ed047575dc34eb793025dec42a500c114f36433cbf1c6e2c3ba06ff"} Dec 03 18:19:07 crc kubenswrapper[4758]: I1203 18:19:07.042672 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3","Type":"ContainerStarted","Data":"ad3b8b65942b562d09baabf386ccf9fc4a2397e4e5c23ad66f67af4ccd0a0291"} Dec 03 18:19:07 crc kubenswrapper[4758]: I1203 18:19:07.092652 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.092632183 podStartE2EDuration="8.092632183s" podCreationTimestamp="2025-12-03 18:18:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:07.084567515 +0000 UTC m=+5002.285944376" watchObservedRunningTime="2025-12-03 18:19:07.092632183 +0000 UTC m=+5002.294009044" Dec 03 18:19:08 crc kubenswrapper[4758]: I1203 18:19:08.050371 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0bb80e50-176a-4fc0-890d-026294efeaaa","Type":"ContainerStarted","Data":"5f299433e3ca47eb8bb89d7e8095399932f51a754940652e17af31989b757bb0"} Dec 03 18:19:08 crc kubenswrapper[4758]: I1203 18:19:08.071047 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.071013219 podStartE2EDuration="8.071013219s" podCreationTimestamp="2025-12-03 18:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:08.066452196 +0000 UTC m=+5003.267829057" watchObservedRunningTime="2025-12-03 18:19:08.071013219 +0000 UTC m=+5003.272390080" Dec 03 18:19:08 crc kubenswrapper[4758]: I1203 18:19:08.092845 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:19:08 crc kubenswrapper[4758]: I1203 18:19:08.298754 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:19:08 crc kubenswrapper[4758]: I1203 18:19:08.344146 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-5nlxk"] Dec 03 18:19:09 crc kubenswrapper[4758]: I1203 18:19:09.056855 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerName="dnsmasq-dns" containerID="cri-o://95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb" gracePeriod=10 Dec 03 18:19:09 crc kubenswrapper[4758]: I1203 18:19:09.975110 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.065834 4758 generic.go:334] "Generic (PLEG): container finished" podID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerID="95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb" exitCode=0 Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.065884 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" event={"ID":"9a35df35-4f43-4720-b12a-b393f9f56c64","Type":"ContainerDied","Data":"95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb"} Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.065910 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" event={"ID":"9a35df35-4f43-4720-b12a-b393f9f56c64","Type":"ContainerDied","Data":"ac16cc7b9d3195aaa32b4fe157573efebc8c87fa472839d01fd62d73ce423bdb"} Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.065926 4758 scope.go:117] "RemoveContainer" containerID="95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.066031 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-5nlxk" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.105840 4758 scope.go:117] "RemoveContainer" containerID="dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.120448 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-config\") pod \"9a35df35-4f43-4720-b12a-b393f9f56c64\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.120568 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-dns-svc\") pod \"9a35df35-4f43-4720-b12a-b393f9f56c64\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.120607 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvh77\" (UniqueName: \"kubernetes.io/projected/9a35df35-4f43-4720-b12a-b393f9f56c64-kube-api-access-fvh77\") pod \"9a35df35-4f43-4720-b12a-b393f9f56c64\" (UID: \"9a35df35-4f43-4720-b12a-b393f9f56c64\") " Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.139931 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a35df35-4f43-4720-b12a-b393f9f56c64-kube-api-access-fvh77" (OuterVolumeSpecName: "kube-api-access-fvh77") pod "9a35df35-4f43-4720-b12a-b393f9f56c64" (UID: "9a35df35-4f43-4720-b12a-b393f9f56c64"). InnerVolumeSpecName "kube-api-access-fvh77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.169454 4758 scope.go:117] "RemoveContainer" containerID="95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb" Dec 03 18:19:10 crc kubenswrapper[4758]: E1203 18:19:10.170070 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb\": container with ID starting with 95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb not found: ID does not exist" containerID="95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.170109 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb"} err="failed to get container status \"95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb\": rpc error: code = NotFound desc = could not find container \"95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb\": container with ID starting with 95fe8b58c24704032dc2dbbe492ea38feeac00e06a7c79e37d22dfd17e06d0fb not found: ID does not exist" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.170134 4758 scope.go:117] "RemoveContainer" containerID="dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc" Dec 03 18:19:10 crc kubenswrapper[4758]: E1203 18:19:10.170385 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc\": container with ID starting with dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc not found: ID does not exist" containerID="dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.170406 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc"} err="failed to get container status \"dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc\": rpc error: code = NotFound desc = could not find container \"dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc\": container with ID starting with dcca604e528201de4d299250e70baf2c5b225760d65be25313a7ff547447d5bc not found: ID does not exist" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.188935 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a35df35-4f43-4720-b12a-b393f9f56c64" (UID: "9a35df35-4f43-4720-b12a-b393f9f56c64"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.193768 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-config" (OuterVolumeSpecName: "config") pod "9a35df35-4f43-4720-b12a-b393f9f56c64" (UID: "9a35df35-4f43-4720-b12a-b393f9f56c64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.222249 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.222286 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvh77\" (UniqueName: \"kubernetes.io/projected/9a35df35-4f43-4720-b12a-b393f9f56c64-kube-api-access-fvh77\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.222299 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a35df35-4f43-4720-b12a-b393f9f56c64-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.410029 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-5nlxk"] Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.422884 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-5nlxk"] Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.684038 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.684082 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 18:19:10 crc kubenswrapper[4758]: I1203 18:19:10.747364 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 18:19:11 crc kubenswrapper[4758]: I1203 18:19:11.122908 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" path="/var/lib/kubelet/pods/9a35df35-4f43-4720-b12a-b393f9f56c64/volumes" Dec 03 18:19:11 crc kubenswrapper[4758]: I1203 18:19:11.135535 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 18:19:12 crc kubenswrapper[4758]: I1203 18:19:12.151278 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:12 crc kubenswrapper[4758]: I1203 18:19:12.151345 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:14 crc kubenswrapper[4758]: I1203 18:19:14.347006 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:14 crc kubenswrapper[4758]: I1203 18:19:14.440874 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 18:19:33 crc kubenswrapper[4758]: I1203 18:19:33.237538 4758 generic.go:334] "Generic (PLEG): container finished" podID="386626ba-ab82-4063-880e-96518ca1426a" containerID="dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56" exitCode=0 Dec 03 18:19:33 crc kubenswrapper[4758]: I1203 18:19:33.237800 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"386626ba-ab82-4063-880e-96518ca1426a","Type":"ContainerDied","Data":"dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56"} Dec 03 18:19:33 crc kubenswrapper[4758]: I1203 18:19:33.240338 4758 generic.go:334] "Generic (PLEG): container finished" podID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerID="5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e" exitCode=0 Dec 03 18:19:33 crc kubenswrapper[4758]: I1203 18:19:33.240369 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"165e9da9-68ad-45ab-b162-9e70cea488d2","Type":"ContainerDied","Data":"5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e"} Dec 03 18:19:34 crc kubenswrapper[4758]: I1203 18:19:34.249971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"386626ba-ab82-4063-880e-96518ca1426a","Type":"ContainerStarted","Data":"f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42"} Dec 03 18:19:34 crc kubenswrapper[4758]: I1203 18:19:34.252084 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"165e9da9-68ad-45ab-b162-9e70cea488d2","Type":"ContainerStarted","Data":"52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7"} Dec 03 18:19:34 crc kubenswrapper[4758]: I1203 18:19:34.252365 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:19:34 crc kubenswrapper[4758]: I1203 18:19:34.277091 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.277065253 podStartE2EDuration="37.277065253s" podCreationTimestamp="2025-12-03 18:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:34.269724975 +0000 UTC m=+5029.471101856" watchObservedRunningTime="2025-12-03 18:19:34.277065253 +0000 UTC m=+5029.478442114" Dec 03 18:19:34 crc kubenswrapper[4758]: I1203 18:19:34.305605 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.305561032 podStartE2EDuration="36.305561032s" podCreationTimestamp="2025-12-03 18:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:34.299149919 +0000 UTC m=+5029.500526800" watchObservedRunningTime="2025-12-03 18:19:34.305561032 +0000 UTC m=+5029.506937893" Dec 03 18:19:39 crc kubenswrapper[4758]: I1203 18:19:39.495186 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 18:19:41 crc kubenswrapper[4758]: I1203 18:19:41.394869 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:19:41 crc kubenswrapper[4758]: I1203 18:19:41.395251 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:19:49 crc kubenswrapper[4758]: I1203 18:19:49.496613 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 18:19:49 crc kubenswrapper[4758]: I1203 18:19:49.559855 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.022599 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-tsjb2"] Dec 03 18:19:56 crc kubenswrapper[4758]: E1203 18:19:56.023439 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerName="init" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.023454 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerName="init" Dec 03 18:19:56 crc kubenswrapper[4758]: E1203 18:19:56.023475 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerName="dnsmasq-dns" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.023482 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerName="dnsmasq-dns" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.023633 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a35df35-4f43-4720-b12a-b393f9f56c64" containerName="dnsmasq-dns" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.024436 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.043371 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-tsjb2"] Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.161188 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.161257 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62nk\" (UniqueName: \"kubernetes.io/projected/366ef7ee-bf1d-4570-b3c0-da98dccf2685-kube-api-access-f62nk\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.161288 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-config\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.262880 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-config\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.263226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.263368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62nk\" (UniqueName: \"kubernetes.io/projected/366ef7ee-bf1d-4570-b3c0-da98dccf2685-kube-api-access-f62nk\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.264279 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.264466 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-config\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.284941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62nk\" (UniqueName: \"kubernetes.io/projected/366ef7ee-bf1d-4570-b3c0-da98dccf2685-kube-api-access-f62nk\") pod \"dnsmasq-dns-5b7946d7b9-tsjb2\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.342665 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.817406 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-tsjb2"] Dec 03 18:19:56 crc kubenswrapper[4758]: I1203 18:19:56.857551 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:19:57 crc kubenswrapper[4758]: I1203 18:19:57.394976 4758 generic.go:334] "Generic (PLEG): container finished" podID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerID="bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014" exitCode=0 Dec 03 18:19:57 crc kubenswrapper[4758]: I1203 18:19:57.395032 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" event={"ID":"366ef7ee-bf1d-4570-b3c0-da98dccf2685","Type":"ContainerDied","Data":"bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014"} Dec 03 18:19:57 crc kubenswrapper[4758]: I1203 18:19:57.395234 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" event={"ID":"366ef7ee-bf1d-4570-b3c0-da98dccf2685","Type":"ContainerStarted","Data":"053099eff7aa2c0180bee02dd4a43d0898cc9d1a44cbe6d90ae2094a2698f30d"} Dec 03 18:19:57 crc kubenswrapper[4758]: I1203 18:19:57.591057 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:19:58 crc kubenswrapper[4758]: I1203 18:19:58.404746 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" event={"ID":"366ef7ee-bf1d-4570-b3c0-da98dccf2685","Type":"ContainerStarted","Data":"fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6"} Dec 03 18:19:58 crc kubenswrapper[4758]: I1203 18:19:58.405753 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:19:58 crc kubenswrapper[4758]: I1203 18:19:58.425227 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" podStartSLOduration=2.425206783 podStartE2EDuration="2.425206783s" podCreationTimestamp="2025-12-03 18:19:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:19:58.421923174 +0000 UTC m=+5053.623300035" watchObservedRunningTime="2025-12-03 18:19:58.425206783 +0000 UTC m=+5053.626583644" Dec 03 18:19:58 crc kubenswrapper[4758]: I1203 18:19:58.610445 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="rabbitmq" containerID="cri-o://f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42" gracePeriod=604799 Dec 03 18:19:59 crc kubenswrapper[4758]: I1203 18:19:59.394038 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="rabbitmq" containerID="cri-o://52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7" gracePeriod=604799 Dec 03 18:19:59 crc kubenswrapper[4758]: I1203 18:19:59.496352 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.242:5672: connect: connection refused" Dec 03 18:19:59 crc kubenswrapper[4758]: I1203 18:19:59.558648 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.243:5672: connect: connection refused" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.160329 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212269 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/386626ba-ab82-4063-880e-96518ca1426a-pod-info\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212308 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-server-conf\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-plugins\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212367 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-plugins-conf\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212604 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-confd\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212705 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45fdn\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-kube-api-access-45fdn\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212738 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-erlang-cookie\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.212759 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/386626ba-ab82-4063-880e-96518ca1426a-erlang-cookie-secret\") pod \"386626ba-ab82-4063-880e-96518ca1426a\" (UID: \"386626ba-ab82-4063-880e-96518ca1426a\") " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.214051 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.214494 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.214791 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.218633 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386626ba-ab82-4063-880e-96518ca1426a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.225029 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/386626ba-ab82-4063-880e-96518ca1426a-pod-info" (OuterVolumeSpecName: "pod-info") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.225030 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-kube-api-access-45fdn" (OuterVolumeSpecName: "kube-api-access-45fdn") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "kube-api-access-45fdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.227951 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081" (OuterVolumeSpecName: "persistence") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.244416 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-server-conf" (OuterVolumeSpecName: "server-conf") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.294608 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "386626ba-ab82-4063-880e-96518ca1426a" (UID: "386626ba-ab82-4063-880e-96518ca1426a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.314550 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45fdn\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-kube-api-access-45fdn\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.314968 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315036 4758 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/386626ba-ab82-4063-880e-96518ca1426a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315103 4758 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/386626ba-ab82-4063-880e-96518ca1426a-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315161 4758 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315247 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315309 4758 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/386626ba-ab82-4063-880e-96518ca1426a-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315414 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") on node \"crc\" " Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.315485 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/386626ba-ab82-4063-880e-96518ca1426a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.335589 4758 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.335891 4758 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081") on node "crc" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.417319 4758 reconciler_common.go:293] "Volume detached for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.459957 4758 generic.go:334] "Generic (PLEG): container finished" podID="386626ba-ab82-4063-880e-96518ca1426a" containerID="f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42" exitCode=0 Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.459997 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"386626ba-ab82-4063-880e-96518ca1426a","Type":"ContainerDied","Data":"f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42"} Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.460018 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.460026 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"386626ba-ab82-4063-880e-96518ca1426a","Type":"ContainerDied","Data":"4200e392eea3cb229dd00a930df7bcc461fa16f9d7edb195508c653575de031c"} Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.460061 4758 scope.go:117] "RemoveContainer" containerID="f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.501389 4758 scope.go:117] "RemoveContainer" containerID="dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.501547 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.520143 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.527371 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:20:05 crc kubenswrapper[4758]: E1203 18:20:05.527892 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="rabbitmq" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.527921 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="rabbitmq" Dec 03 18:20:05 crc kubenswrapper[4758]: E1203 18:20:05.527935 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="setup-container" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.527946 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="setup-container" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.528134 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="386626ba-ab82-4063-880e-96518ca1426a" containerName="rabbitmq" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.530342 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.535100 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.535414 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.535574 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.535869 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sb5gc" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.536020 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.541413 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.543323 4758 scope.go:117] "RemoveContainer" containerID="f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42" Dec 03 18:20:05 crc kubenswrapper[4758]: E1203 18:20:05.544283 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42\": container with ID starting with f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42 not found: ID does not exist" containerID="f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.544320 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42"} err="failed to get container status \"f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42\": rpc error: code = NotFound desc = could not find container \"f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42\": container with ID starting with f8d831b00b110fb8f6395982ea4aa1362bece69512e567023004cba10e284b42 not found: ID does not exist" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.544342 4758 scope.go:117] "RemoveContainer" containerID="dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56" Dec 03 18:20:05 crc kubenswrapper[4758]: E1203 18:20:05.544606 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56\": container with ID starting with dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56 not found: ID does not exist" containerID="dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.544633 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56"} err="failed to get container status \"dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56\": rpc error: code = NotFound desc = could not find container \"dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56\": container with ID starting with dd4ccdf51f5edbcd11e0ce32b90e0222c083b03dc9f08b386c719314a7c90c56 not found: ID does not exist" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620020 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mkn7\" (UniqueName: \"kubernetes.io/projected/4a99f005-65b5-4d3b-a49f-75213f40b232-kube-api-access-6mkn7\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620433 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620571 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a99f005-65b5-4d3b-a49f-75213f40b232-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620627 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a99f005-65b5-4d3b-a49f-75213f40b232-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620651 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620765 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620816 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a99f005-65b5-4d3b-a49f-75213f40b232-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.620869 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a99f005-65b5-4d3b-a49f-75213f40b232-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722256 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mkn7\" (UniqueName: \"kubernetes.io/projected/4a99f005-65b5-4d3b-a49f-75213f40b232-kube-api-access-6mkn7\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722354 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722400 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a99f005-65b5-4d3b-a49f-75213f40b232-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a99f005-65b5-4d3b-a49f-75213f40b232-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722448 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722482 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722509 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a99f005-65b5-4d3b-a49f-75213f40b232-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722550 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a99f005-65b5-4d3b-a49f-75213f40b232-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722944 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.722943 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.723736 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a99f005-65b5-4d3b-a49f-75213f40b232-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.724065 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a99f005-65b5-4d3b-a49f-75213f40b232-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.726916 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a99f005-65b5-4d3b-a49f-75213f40b232-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.728584 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.728619 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4213de258debccea79f1d56b24d3aacf86159bcd6ca8adb0d34458dcf4f780d4/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.729097 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a99f005-65b5-4d3b-a49f-75213f40b232-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.729456 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a99f005-65b5-4d3b-a49f-75213f40b232-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.738206 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mkn7\" (UniqueName: \"kubernetes.io/projected/4a99f005-65b5-4d3b-a49f-75213f40b232-kube-api-access-6mkn7\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.760848 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e3498692-fe2c-4d61-b02a-f63ca1b32081\") pod \"rabbitmq-server-0\" (UID: \"4a99f005-65b5-4d3b-a49f-75213f40b232\") " pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.898020 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 18:20:05 crc kubenswrapper[4758]: I1203 18:20:05.998529 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028499 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/165e9da9-68ad-45ab-b162-9e70cea488d2-pod-info\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028561 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-plugins-conf\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-erlang-cookie\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028699 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/165e9da9-68ad-45ab-b162-9e70cea488d2-erlang-cookie-secret\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028730 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-confd\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028859 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028943 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dnwz\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-kube-api-access-6dnwz\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.028973 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-server-conf\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.029008 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-plugins\") pod \"165e9da9-68ad-45ab-b162-9e70cea488d2\" (UID: \"165e9da9-68ad-45ab-b162-9e70cea488d2\") " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.030897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.031310 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.032041 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.038515 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/165e9da9-68ad-45ab-b162-9e70cea488d2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.040916 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/165e9da9-68ad-45ab-b162-9e70cea488d2-pod-info" (OuterVolumeSpecName: "pod-info") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.048370 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-kube-api-access-6dnwz" (OuterVolumeSpecName: "kube-api-access-6dnwz") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "kube-api-access-6dnwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.056161 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634" (OuterVolumeSpecName: "persistence") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.060374 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-server-conf" (OuterVolumeSpecName: "server-conf") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131088 4758 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/165e9da9-68ad-45ab-b162-9e70cea488d2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131181 4758 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131203 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131220 4758 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/165e9da9-68ad-45ab-b162-9e70cea488d2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131262 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") on node \"crc\" " Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131287 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dnwz\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-kube-api-access-6dnwz\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131308 4758 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/165e9da9-68ad-45ab-b162-9e70cea488d2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.131325 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.132993 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "165e9da9-68ad-45ab-b162-9e70cea488d2" (UID: "165e9da9-68ad-45ab-b162-9e70cea488d2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.159130 4758 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.159299 4758 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634") on node "crc" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.233027 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/165e9da9-68ad-45ab-b162-9e70cea488d2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.233055 4758 reconciler_common.go:293] "Volume detached for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.345383 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.398241 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.420575 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-t7h6f"] Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.421397 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerName="dnsmasq-dns" containerID="cri-o://1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180" gracePeriod=10 Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.478827 4758 generic.go:334] "Generic (PLEG): container finished" podID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerID="52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7" exitCode=0 Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.479169 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.479898 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"165e9da9-68ad-45ab-b162-9e70cea488d2","Type":"ContainerDied","Data":"52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7"} Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.479953 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"165e9da9-68ad-45ab-b162-9e70cea488d2","Type":"ContainerDied","Data":"0da5450239decef359be067d28da09fb1aa702bb8a0e3490bfe073eef3bc02a6"} Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.479973 4758 scope.go:117] "RemoveContainer" containerID="52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.485114 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a99f005-65b5-4d3b-a49f-75213f40b232","Type":"ContainerStarted","Data":"188863ae352f970bd471803d18b5ba2020702f9343ba312f316eaa7e2c47f339"} Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.501624 4758 scope.go:117] "RemoveContainer" containerID="5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.523098 4758 scope.go:117] "RemoveContainer" containerID="52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7" Dec 03 18:20:06 crc kubenswrapper[4758]: E1203 18:20:06.523713 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7\": container with ID starting with 52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7 not found: ID does not exist" containerID="52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.523743 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7"} err="failed to get container status \"52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7\": rpc error: code = NotFound desc = could not find container \"52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7\": container with ID starting with 52fdf3ee3b256e90170c5ec039c99f8b7b442d7479758d2c82e7a7f3f479a8b7 not found: ID does not exist" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.523764 4758 scope.go:117] "RemoveContainer" containerID="5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e" Dec 03 18:20:06 crc kubenswrapper[4758]: E1203 18:20:06.524094 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e\": container with ID starting with 5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e not found: ID does not exist" containerID="5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.524112 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e"} err="failed to get container status \"5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e\": rpc error: code = NotFound desc = could not find container \"5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e\": container with ID starting with 5e3b337b7503be4cd7ffda723e3ead75c41253aff58d31508a0ac9788872460e not found: ID does not exist" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.524967 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.529267 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.557393 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:20:06 crc kubenswrapper[4758]: E1203 18:20:06.557789 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="setup-container" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.557805 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="setup-container" Dec 03 18:20:06 crc kubenswrapper[4758]: E1203 18:20:06.557822 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="rabbitmq" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.557829 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="rabbitmq" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.557979 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" containerName="rabbitmq" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.558792 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.564652 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.566714 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.567016 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.567046 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7xm5w" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.567070 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.567202 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643487 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6707c0c8-b302-46bf-a6d7-e953f5f2d497-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643549 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6707c0c8-b302-46bf-a6d7-e953f5f2d497-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643604 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643656 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643699 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643724 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5ngt\" (UniqueName: \"kubernetes.io/projected/6707c0c8-b302-46bf-a6d7-e953f5f2d497-kube-api-access-d5ngt\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643760 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6707c0c8-b302-46bf-a6d7-e953f5f2d497-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643794 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6707c0c8-b302-46bf-a6d7-e953f5f2d497-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:06 crc kubenswrapper[4758]: I1203 18:20:06.643815 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6707c0c8-b302-46bf-a6d7-e953f5f2d497-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745073 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6707c0c8-b302-46bf-a6d7-e953f5f2d497-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745114 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745149 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745169 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745186 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5ngt\" (UniqueName: \"kubernetes.io/projected/6707c0c8-b302-46bf-a6d7-e953f5f2d497-kube-api-access-d5ngt\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6707c0c8-b302-46bf-a6d7-e953f5f2d497-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6707c0c8-b302-46bf-a6d7-e953f5f2d497-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.745287 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.754824 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.755248 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.755852 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6707c0c8-b302-46bf-a6d7-e953f5f2d497-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.756285 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6707c0c8-b302-46bf-a6d7-e953f5f2d497-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.757629 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6707c0c8-b302-46bf-a6d7-e953f5f2d497-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.758020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6707c0c8-b302-46bf-a6d7-e953f5f2d497-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.783350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6707c0c8-b302-46bf-a6d7-e953f5f2d497-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.799235 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5ngt\" (UniqueName: \"kubernetes.io/projected/6707c0c8-b302-46bf-a6d7-e953f5f2d497-kube-api-access-d5ngt\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.801965 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.802000 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2613edf6caccdc4198d192ba5263c37a33bb7fb662d0fa7d899fb5d156d73312/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.842116 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f6abe33-081f-4fb6-abc4-af674c1f3634\") pod \"rabbitmq-cell1-server-0\" (UID: \"6707c0c8-b302-46bf-a6d7-e953f5f2d497\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:06.924013 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.124173 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="165e9da9-68ad-45ab-b162-9e70cea488d2" path="/var/lib/kubelet/pods/165e9da9-68ad-45ab-b162-9e70cea488d2/volumes" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.125405 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="386626ba-ab82-4063-880e-96518ca1426a" path="/var/lib/kubelet/pods/386626ba-ab82-4063-880e-96518ca1426a/volumes" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.190363 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.252947 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfnhc\" (UniqueName: \"kubernetes.io/projected/073714d5-1c99-44bf-8bf2-d14dc3a08e32-kube-api-access-bfnhc\") pod \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.253140 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-dns-svc\") pod \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.253210 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-config\") pod \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\" (UID: \"073714d5-1c99-44bf-8bf2-d14dc3a08e32\") " Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.333161 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073714d5-1c99-44bf-8bf2-d14dc3a08e32-kube-api-access-bfnhc" (OuterVolumeSpecName: "kube-api-access-bfnhc") pod "073714d5-1c99-44bf-8bf2-d14dc3a08e32" (UID: "073714d5-1c99-44bf-8bf2-d14dc3a08e32"). InnerVolumeSpecName "kube-api-access-bfnhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.354921 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfnhc\" (UniqueName: \"kubernetes.io/projected/073714d5-1c99-44bf-8bf2-d14dc3a08e32-kube-api-access-bfnhc\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.362014 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-config" (OuterVolumeSpecName: "config") pod "073714d5-1c99-44bf-8bf2-d14dc3a08e32" (UID: "073714d5-1c99-44bf-8bf2-d14dc3a08e32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.374585 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "073714d5-1c99-44bf-8bf2-d14dc3a08e32" (UID: "073714d5-1c99-44bf-8bf2-d14dc3a08e32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.456731 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.456755 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073714d5-1c99-44bf-8bf2-d14dc3a08e32-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.500653 4758 generic.go:334] "Generic (PLEG): container finished" podID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerID="1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180" exitCode=0 Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.500738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" event={"ID":"073714d5-1c99-44bf-8bf2-d14dc3a08e32","Type":"ContainerDied","Data":"1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180"} Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.500752 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.500777 4758 scope.go:117] "RemoveContainer" containerID="1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.500763 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-t7h6f" event={"ID":"073714d5-1c99-44bf-8bf2-d14dc3a08e32","Type":"ContainerDied","Data":"3f1d509dc49917f6cf3d47d600402fc5798c31d035dfb82ef8200a75d0f20a90"} Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.510856 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a99f005-65b5-4d3b-a49f-75213f40b232","Type":"ContainerStarted","Data":"2a30419c98f02afe4a041e9002f45ecd3ba90f12c7f8a2988e125ab98480c761"} Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.551382 4758 scope.go:117] "RemoveContainer" containerID="fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.577842 4758 scope.go:117] "RemoveContainer" containerID="1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180" Dec 03 18:20:07 crc kubenswrapper[4758]: E1203 18:20:07.578326 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180\": container with ID starting with 1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180 not found: ID does not exist" containerID="1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.578355 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180"} err="failed to get container status \"1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180\": rpc error: code = NotFound desc = could not find container \"1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180\": container with ID starting with 1e2a4128cefe3851e81115a59ef69cab5187e846ac63b1efc4835a5951c27180 not found: ID does not exist" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.578375 4758 scope.go:117] "RemoveContainer" containerID="fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa" Dec 03 18:20:07 crc kubenswrapper[4758]: E1203 18:20:07.578578 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa\": container with ID starting with fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa not found: ID does not exist" containerID="fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.578594 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa"} err="failed to get container status \"fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa\": rpc error: code = NotFound desc = could not find container \"fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa\": container with ID starting with fcfc4aac99fddddf92497c449de5dce805749a91724e539263cdfd091a3e90aa not found: ID does not exist" Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.579826 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-t7h6f"] Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.594952 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-t7h6f"] Dec 03 18:20:07 crc kubenswrapper[4758]: I1203 18:20:07.748660 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 18:20:07 crc kubenswrapper[4758]: W1203 18:20:07.751353 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6707c0c8_b302_46bf_a6d7_e953f5f2d497.slice/crio-f5f4423ac77da5b8f2a5f5381bbe64e228dcfeed21831774d1aea3af5330d3f4 WatchSource:0}: Error finding container f5f4423ac77da5b8f2a5f5381bbe64e228dcfeed21831774d1aea3af5330d3f4: Status 404 returned error can't find the container with id f5f4423ac77da5b8f2a5f5381bbe64e228dcfeed21831774d1aea3af5330d3f4 Dec 03 18:20:08 crc kubenswrapper[4758]: I1203 18:20:08.520320 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6707c0c8-b302-46bf-a6d7-e953f5f2d497","Type":"ContainerStarted","Data":"f5f4423ac77da5b8f2a5f5381bbe64e228dcfeed21831774d1aea3af5330d3f4"} Dec 03 18:20:09 crc kubenswrapper[4758]: I1203 18:20:09.125280 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" path="/var/lib/kubelet/pods/073714d5-1c99-44bf-8bf2-d14dc3a08e32/volumes" Dec 03 18:20:09 crc kubenswrapper[4758]: I1203 18:20:09.529389 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6707c0c8-b302-46bf-a6d7-e953f5f2d497","Type":"ContainerStarted","Data":"fb7de86963088d4f48e8b986eb0e5d181430cf94d66241d826f4cf1f0276483d"} Dec 03 18:20:11 crc kubenswrapper[4758]: I1203 18:20:11.394662 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:20:11 crc kubenswrapper[4758]: I1203 18:20:11.395072 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:20:39 crc kubenswrapper[4758]: I1203 18:20:39.774726 4758 generic.go:334] "Generic (PLEG): container finished" podID="4a99f005-65b5-4d3b-a49f-75213f40b232" containerID="2a30419c98f02afe4a041e9002f45ecd3ba90f12c7f8a2988e125ab98480c761" exitCode=0 Dec 03 18:20:39 crc kubenswrapper[4758]: I1203 18:20:39.774820 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a99f005-65b5-4d3b-a49f-75213f40b232","Type":"ContainerDied","Data":"2a30419c98f02afe4a041e9002f45ecd3ba90f12c7f8a2988e125ab98480c761"} Dec 03 18:20:40 crc kubenswrapper[4758]: I1203 18:20:40.788352 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a99f005-65b5-4d3b-a49f-75213f40b232","Type":"ContainerStarted","Data":"06e5739104d1daebead0e6c013d04cee76e1df040dd26594f864572ae4222957"} Dec 03 18:20:40 crc kubenswrapper[4758]: I1203 18:20:40.789289 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 18:20:40 crc kubenswrapper[4758]: I1203 18:20:40.828034 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.828006728 podStartE2EDuration="35.828006728s" podCreationTimestamp="2025-12-03 18:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:20:40.813049855 +0000 UTC m=+5096.014426746" watchObservedRunningTime="2025-12-03 18:20:40.828006728 +0000 UTC m=+5096.029383599" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.395417 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.395533 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.395599 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.396414 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.396509 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" gracePeriod=600 Dec 03 18:20:41 crc kubenswrapper[4758]: E1203 18:20:41.522298 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.803180 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" exitCode=0 Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.803267 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176"} Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.803360 4758 scope.go:117] "RemoveContainer" containerID="9807157ac08bd499c3c2865cca4e64a21569ebe37d2458c786adcfa7ed1dcacf" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.804216 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:20:41 crc kubenswrapper[4758]: E1203 18:20:41.804630 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.806539 4758 generic.go:334] "Generic (PLEG): container finished" podID="6707c0c8-b302-46bf-a6d7-e953f5f2d497" containerID="fb7de86963088d4f48e8b986eb0e5d181430cf94d66241d826f4cf1f0276483d" exitCode=0 Dec 03 18:20:41 crc kubenswrapper[4758]: I1203 18:20:41.806619 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6707c0c8-b302-46bf-a6d7-e953f5f2d497","Type":"ContainerDied","Data":"fb7de86963088d4f48e8b986eb0e5d181430cf94d66241d826f4cf1f0276483d"} Dec 03 18:20:42 crc kubenswrapper[4758]: I1203 18:20:42.819545 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6707c0c8-b302-46bf-a6d7-e953f5f2d497","Type":"ContainerStarted","Data":"5f964b64cb3c5c586838c2742b75dc8e0f85ceb9b7b8ebab72ae349dae6b471b"} Dec 03 18:20:42 crc kubenswrapper[4758]: I1203 18:20:42.820507 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:20:42 crc kubenswrapper[4758]: I1203 18:20:42.854096 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.854072073 podStartE2EDuration="36.854072073s" podCreationTimestamp="2025-12-03 18:20:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:20:42.849279683 +0000 UTC m=+5098.050656564" watchObservedRunningTime="2025-12-03 18:20:42.854072073 +0000 UTC m=+5098.055448934" Dec 03 18:20:55 crc kubenswrapper[4758]: I1203 18:20:55.900777 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 18:20:56 crc kubenswrapper[4758]: I1203 18:20:56.114295 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:20:56 crc kubenswrapper[4758]: E1203 18:20:56.114896 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:20:56 crc kubenswrapper[4758]: I1203 18:20:56.927868 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.529156 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 03 18:21:04 crc kubenswrapper[4758]: E1203 18:21:04.530049 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerName="init" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.530068 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerName="init" Dec 03 18:21:04 crc kubenswrapper[4758]: E1203 18:21:04.530100 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerName="dnsmasq-dns" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.530108 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerName="dnsmasq-dns" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.530274 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="073714d5-1c99-44bf-8bf2-d14dc3a08e32" containerName="dnsmasq-dns" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.530956 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.532792 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrxtm" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.540128 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.641756 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8svh\" (UniqueName: \"kubernetes.io/projected/36ef9526-7e9b-4f2b-9ab5-ff8db1853450-kube-api-access-r8svh\") pod \"mariadb-client-1-default\" (UID: \"36ef9526-7e9b-4f2b-9ab5-ff8db1853450\") " pod="openstack/mariadb-client-1-default" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.742768 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8svh\" (UniqueName: \"kubernetes.io/projected/36ef9526-7e9b-4f2b-9ab5-ff8db1853450-kube-api-access-r8svh\") pod \"mariadb-client-1-default\" (UID: \"36ef9526-7e9b-4f2b-9ab5-ff8db1853450\") " pod="openstack/mariadb-client-1-default" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.761559 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8svh\" (UniqueName: \"kubernetes.io/projected/36ef9526-7e9b-4f2b-9ab5-ff8db1853450-kube-api-access-r8svh\") pod \"mariadb-client-1-default\" (UID: \"36ef9526-7e9b-4f2b-9ab5-ff8db1853450\") " pod="openstack/mariadb-client-1-default" Dec 03 18:21:04 crc kubenswrapper[4758]: I1203 18:21:04.885022 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 03 18:21:05 crc kubenswrapper[4758]: I1203 18:21:05.380578 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 03 18:21:06 crc kubenswrapper[4758]: I1203 18:21:06.009225 4758 generic.go:334] "Generic (PLEG): container finished" podID="36ef9526-7e9b-4f2b-9ab5-ff8db1853450" containerID="71b523946e959456b2486c94a4a693522116bad2293b7b72081619f367467920" exitCode=0 Dec 03 18:21:06 crc kubenswrapper[4758]: I1203 18:21:06.009545 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"36ef9526-7e9b-4f2b-9ab5-ff8db1853450","Type":"ContainerDied","Data":"71b523946e959456b2486c94a4a693522116bad2293b7b72081619f367467920"} Dec 03 18:21:06 crc kubenswrapper[4758]: I1203 18:21:06.009601 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"36ef9526-7e9b-4f2b-9ab5-ff8db1853450","Type":"ContainerStarted","Data":"c16baeaa7abbbe1768c80851406f1ef8fa83ccbb050bfb8367430b4963454f70"} Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.433893 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.463737 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_36ef9526-7e9b-4f2b-9ab5-ff8db1853450/mariadb-client-1-default/0.log" Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.498847 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.513140 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.584055 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8svh\" (UniqueName: \"kubernetes.io/projected/36ef9526-7e9b-4f2b-9ab5-ff8db1853450-kube-api-access-r8svh\") pod \"36ef9526-7e9b-4f2b-9ab5-ff8db1853450\" (UID: \"36ef9526-7e9b-4f2b-9ab5-ff8db1853450\") " Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.588816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36ef9526-7e9b-4f2b-9ab5-ff8db1853450-kube-api-access-r8svh" (OuterVolumeSpecName: "kube-api-access-r8svh") pod "36ef9526-7e9b-4f2b-9ab5-ff8db1853450" (UID: "36ef9526-7e9b-4f2b-9ab5-ff8db1853450"). InnerVolumeSpecName "kube-api-access-r8svh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:07 crc kubenswrapper[4758]: I1203 18:21:07.685479 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8svh\" (UniqueName: \"kubernetes.io/projected/36ef9526-7e9b-4f2b-9ab5-ff8db1853450-kube-api-access-r8svh\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.024577 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c16baeaa7abbbe1768c80851406f1ef8fa83ccbb050bfb8367430b4963454f70" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.024671 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.056270 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 03 18:21:08 crc kubenswrapper[4758]: E1203 18:21:08.057022 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36ef9526-7e9b-4f2b-9ab5-ff8db1853450" containerName="mariadb-client-1-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.057048 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="36ef9526-7e9b-4f2b-9ab5-ff8db1853450" containerName="mariadb-client-1-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.057279 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="36ef9526-7e9b-4f2b-9ab5-ff8db1853450" containerName="mariadb-client-1-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.057974 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.060850 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrxtm" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.061822 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.193442 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67hb2\" (UniqueName: \"kubernetes.io/projected/daebe10e-6f0f-4c32-849f-955db3299533-kube-api-access-67hb2\") pod \"mariadb-client-2-default\" (UID: \"daebe10e-6f0f-4c32-849f-955db3299533\") " pod="openstack/mariadb-client-2-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.295156 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67hb2\" (UniqueName: \"kubernetes.io/projected/daebe10e-6f0f-4c32-849f-955db3299533-kube-api-access-67hb2\") pod \"mariadb-client-2-default\" (UID: \"daebe10e-6f0f-4c32-849f-955db3299533\") " pod="openstack/mariadb-client-2-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.311932 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67hb2\" (UniqueName: \"kubernetes.io/projected/daebe10e-6f0f-4c32-849f-955db3299533-kube-api-access-67hb2\") pod \"mariadb-client-2-default\" (UID: \"daebe10e-6f0f-4c32-849f-955db3299533\") " pod="openstack/mariadb-client-2-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.375624 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 03 18:21:08 crc kubenswrapper[4758]: I1203 18:21:08.895543 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 03 18:21:09 crc kubenswrapper[4758]: I1203 18:21:09.033659 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"daebe10e-6f0f-4c32-849f-955db3299533","Type":"ContainerStarted","Data":"db3ae142fdbea6454a6d5faec1aa5219620c2679e2a084b60844976335974446"} Dec 03 18:21:09 crc kubenswrapper[4758]: I1203 18:21:09.124411 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36ef9526-7e9b-4f2b-9ab5-ff8db1853450" path="/var/lib/kubelet/pods/36ef9526-7e9b-4f2b-9ab5-ff8db1853450/volumes" Dec 03 18:21:10 crc kubenswrapper[4758]: I1203 18:21:10.044336 4758 generic.go:334] "Generic (PLEG): container finished" podID="daebe10e-6f0f-4c32-849f-955db3299533" containerID="1d2f671bd647a86d14ca5e8bfec6b0b60d9535f422bafd64df030c37f2816ed2" exitCode=1 Dec 03 18:21:10 crc kubenswrapper[4758]: I1203 18:21:10.044540 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"daebe10e-6f0f-4c32-849f-955db3299533","Type":"ContainerDied","Data":"1d2f671bd647a86d14ca5e8bfec6b0b60d9535f422bafd64df030c37f2816ed2"} Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.114933 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:21:11 crc kubenswrapper[4758]: E1203 18:21:11.115543 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.461748 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.484333 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_daebe10e-6f0f-4c32-849f-955db3299533/mariadb-client-2-default/0.log" Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.517913 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.523592 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.647943 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67hb2\" (UniqueName: \"kubernetes.io/projected/daebe10e-6f0f-4c32-849f-955db3299533-kube-api-access-67hb2\") pod \"daebe10e-6f0f-4c32-849f-955db3299533\" (UID: \"daebe10e-6f0f-4c32-849f-955db3299533\") " Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.653412 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daebe10e-6f0f-4c32-849f-955db3299533-kube-api-access-67hb2" (OuterVolumeSpecName: "kube-api-access-67hb2") pod "daebe10e-6f0f-4c32-849f-955db3299533" (UID: "daebe10e-6f0f-4c32-849f-955db3299533"). InnerVolumeSpecName "kube-api-access-67hb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:11 crc kubenswrapper[4758]: I1203 18:21:11.750266 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67hb2\" (UniqueName: \"kubernetes.io/projected/daebe10e-6f0f-4c32-849f-955db3299533-kube-api-access-67hb2\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.041904 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 03 18:21:12 crc kubenswrapper[4758]: E1203 18:21:12.042307 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daebe10e-6f0f-4c32-849f-955db3299533" containerName="mariadb-client-2-default" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.042332 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="daebe10e-6f0f-4c32-849f-955db3299533" containerName="mariadb-client-2-default" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.042554 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="daebe10e-6f0f-4c32-849f-955db3299533" containerName="mariadb-client-2-default" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.043215 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.065509 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db3ae142fdbea6454a6d5faec1aa5219620c2679e2a084b60844976335974446" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.065576 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.070525 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.158430 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl52n\" (UniqueName: \"kubernetes.io/projected/e57805b9-d82e-4a1e-9159-49a51623e524-kube-api-access-tl52n\") pod \"mariadb-client-1\" (UID: \"e57805b9-d82e-4a1e-9159-49a51623e524\") " pod="openstack/mariadb-client-1" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.260785 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl52n\" (UniqueName: \"kubernetes.io/projected/e57805b9-d82e-4a1e-9159-49a51623e524-kube-api-access-tl52n\") pod \"mariadb-client-1\" (UID: \"e57805b9-d82e-4a1e-9159-49a51623e524\") " pod="openstack/mariadb-client-1" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.282432 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl52n\" (UniqueName: \"kubernetes.io/projected/e57805b9-d82e-4a1e-9159-49a51623e524-kube-api-access-tl52n\") pod \"mariadb-client-1\" (UID: \"e57805b9-d82e-4a1e-9159-49a51623e524\") " pod="openstack/mariadb-client-1" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.367115 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 03 18:21:12 crc kubenswrapper[4758]: I1203 18:21:12.881028 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 03 18:21:13 crc kubenswrapper[4758]: I1203 18:21:13.075001 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"e57805b9-d82e-4a1e-9159-49a51623e524","Type":"ContainerStarted","Data":"1f2012aa717cb01efdfb58449db37d2a38e88ade7662d67a229c2be66dd0a8fc"} Dec 03 18:21:13 crc kubenswrapper[4758]: I1203 18:21:13.122755 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daebe10e-6f0f-4c32-849f-955db3299533" path="/var/lib/kubelet/pods/daebe10e-6f0f-4c32-849f-955db3299533/volumes" Dec 03 18:21:14 crc kubenswrapper[4758]: I1203 18:21:14.087811 4758 generic.go:334] "Generic (PLEG): container finished" podID="e57805b9-d82e-4a1e-9159-49a51623e524" containerID="1a0e51cfde4ed5e4594e8cd09e944ec56b1803e593417d1381a7c40d42a3247d" exitCode=0 Dec 03 18:21:14 crc kubenswrapper[4758]: I1203 18:21:14.087923 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"e57805b9-d82e-4a1e-9159-49a51623e524","Type":"ContainerDied","Data":"1a0e51cfde4ed5e4594e8cd09e944ec56b1803e593417d1381a7c40d42a3247d"} Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.449469 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.468668 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_e57805b9-d82e-4a1e-9159-49a51623e524/mariadb-client-1/0.log" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.496891 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.504709 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.609975 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl52n\" (UniqueName: \"kubernetes.io/projected/e57805b9-d82e-4a1e-9159-49a51623e524-kube-api-access-tl52n\") pod \"e57805b9-d82e-4a1e-9159-49a51623e524\" (UID: \"e57805b9-d82e-4a1e-9159-49a51623e524\") " Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.617513 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57805b9-d82e-4a1e-9159-49a51623e524-kube-api-access-tl52n" (OuterVolumeSpecName: "kube-api-access-tl52n") pod "e57805b9-d82e-4a1e-9159-49a51623e524" (UID: "e57805b9-d82e-4a1e-9159-49a51623e524"). InnerVolumeSpecName "kube-api-access-tl52n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.711808 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl52n\" (UniqueName: \"kubernetes.io/projected/e57805b9-d82e-4a1e-9159-49a51623e524-kube-api-access-tl52n\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.893041 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 03 18:21:15 crc kubenswrapper[4758]: E1203 18:21:15.893638 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57805b9-d82e-4a1e-9159-49a51623e524" containerName="mariadb-client-1" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.893764 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57805b9-d82e-4a1e-9159-49a51623e524" containerName="mariadb-client-1" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.894204 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57805b9-d82e-4a1e-9159-49a51623e524" containerName="mariadb-client-1" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.895256 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 03 18:21:15 crc kubenswrapper[4758]: I1203 18:21:15.903921 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.016378 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79l2f\" (UniqueName: \"kubernetes.io/projected/83e567a5-4ba4-45bb-a66a-74298eeb7e48-kube-api-access-79l2f\") pod \"mariadb-client-4-default\" (UID: \"83e567a5-4ba4-45bb-a66a-74298eeb7e48\") " pod="openstack/mariadb-client-4-default" Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.103500 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f2012aa717cb01efdfb58449db37d2a38e88ade7662d67a229c2be66dd0a8fc" Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.103578 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.118969 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79l2f\" (UniqueName: \"kubernetes.io/projected/83e567a5-4ba4-45bb-a66a-74298eeb7e48-kube-api-access-79l2f\") pod \"mariadb-client-4-default\" (UID: \"83e567a5-4ba4-45bb-a66a-74298eeb7e48\") " pod="openstack/mariadb-client-4-default" Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.145204 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79l2f\" (UniqueName: \"kubernetes.io/projected/83e567a5-4ba4-45bb-a66a-74298eeb7e48-kube-api-access-79l2f\") pod \"mariadb-client-4-default\" (UID: \"83e567a5-4ba4-45bb-a66a-74298eeb7e48\") " pod="openstack/mariadb-client-4-default" Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.213749 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 03 18:21:16 crc kubenswrapper[4758]: I1203 18:21:16.736437 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 03 18:21:16 crc kubenswrapper[4758]: W1203 18:21:16.743114 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83e567a5_4ba4_45bb_a66a_74298eeb7e48.slice/crio-69c34ac678c538ec26cde17d8503bf2ebe724e56b2470143815050c48262643d WatchSource:0}: Error finding container 69c34ac678c538ec26cde17d8503bf2ebe724e56b2470143815050c48262643d: Status 404 returned error can't find the container with id 69c34ac678c538ec26cde17d8503bf2ebe724e56b2470143815050c48262643d Dec 03 18:21:17 crc kubenswrapper[4758]: I1203 18:21:17.113384 4758 generic.go:334] "Generic (PLEG): container finished" podID="83e567a5-4ba4-45bb-a66a-74298eeb7e48" containerID="6d80b5d3a98378af07f384d0de9bd25dd1e092834d190cc3653710aa32de1c27" exitCode=0 Dec 03 18:21:17 crc kubenswrapper[4758]: I1203 18:21:17.121866 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e57805b9-d82e-4a1e-9159-49a51623e524" path="/var/lib/kubelet/pods/e57805b9-d82e-4a1e-9159-49a51623e524/volumes" Dec 03 18:21:17 crc kubenswrapper[4758]: I1203 18:21:17.122437 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"83e567a5-4ba4-45bb-a66a-74298eeb7e48","Type":"ContainerDied","Data":"6d80b5d3a98378af07f384d0de9bd25dd1e092834d190cc3653710aa32de1c27"} Dec 03 18:21:17 crc kubenswrapper[4758]: I1203 18:21:17.122514 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"83e567a5-4ba4-45bb-a66a-74298eeb7e48","Type":"ContainerStarted","Data":"69c34ac678c538ec26cde17d8503bf2ebe724e56b2470143815050c48262643d"} Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.521305 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.537136 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_83e567a5-4ba4-45bb-a66a-74298eeb7e48/mariadb-client-4-default/0.log" Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.560817 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.566232 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.662670 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79l2f\" (UniqueName: \"kubernetes.io/projected/83e567a5-4ba4-45bb-a66a-74298eeb7e48-kube-api-access-79l2f\") pod \"83e567a5-4ba4-45bb-a66a-74298eeb7e48\" (UID: \"83e567a5-4ba4-45bb-a66a-74298eeb7e48\") " Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.670656 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e567a5-4ba4-45bb-a66a-74298eeb7e48-kube-api-access-79l2f" (OuterVolumeSpecName: "kube-api-access-79l2f") pod "83e567a5-4ba4-45bb-a66a-74298eeb7e48" (UID: "83e567a5-4ba4-45bb-a66a-74298eeb7e48"). InnerVolumeSpecName "kube-api-access-79l2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:18 crc kubenswrapper[4758]: I1203 18:21:18.764723 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79l2f\" (UniqueName: \"kubernetes.io/projected/83e567a5-4ba4-45bb-a66a-74298eeb7e48-kube-api-access-79l2f\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:19 crc kubenswrapper[4758]: I1203 18:21:19.121863 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e567a5-4ba4-45bb-a66a-74298eeb7e48" path="/var/lib/kubelet/pods/83e567a5-4ba4-45bb-a66a-74298eeb7e48/volumes" Dec 03 18:21:19 crc kubenswrapper[4758]: I1203 18:21:19.127988 4758 scope.go:117] "RemoveContainer" containerID="6d80b5d3a98378af07f384d0de9bd25dd1e092834d190cc3653710aa32de1c27" Dec 03 18:21:19 crc kubenswrapper[4758]: I1203 18:21:19.128141 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.409663 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 03 18:21:22 crc kubenswrapper[4758]: E1203 18:21:22.410898 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e567a5-4ba4-45bb-a66a-74298eeb7e48" containerName="mariadb-client-4-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.410917 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e567a5-4ba4-45bb-a66a-74298eeb7e48" containerName="mariadb-client-4-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.411357 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e567a5-4ba4-45bb-a66a-74298eeb7e48" containerName="mariadb-client-4-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.412285 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.412387 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.416363 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrxtm" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.522136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz2mh\" (UniqueName: \"kubernetes.io/projected/30983465-84ac-482b-b11e-ceb9a2409ad3-kube-api-access-nz2mh\") pod \"mariadb-client-5-default\" (UID: \"30983465-84ac-482b-b11e-ceb9a2409ad3\") " pod="openstack/mariadb-client-5-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.566187 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6pz"] Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.569136 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.575238 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6pz"] Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.623890 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz2mh\" (UniqueName: \"kubernetes.io/projected/30983465-84ac-482b-b11e-ceb9a2409ad3-kube-api-access-nz2mh\") pod \"mariadb-client-5-default\" (UID: \"30983465-84ac-482b-b11e-ceb9a2409ad3\") " pod="openstack/mariadb-client-5-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.644505 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz2mh\" (UniqueName: \"kubernetes.io/projected/30983465-84ac-482b-b11e-ceb9a2409ad3-kube-api-access-nz2mh\") pod \"mariadb-client-5-default\" (UID: \"30983465-84ac-482b-b11e-ceb9a2409ad3\") " pod="openstack/mariadb-client-5-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.725058 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-catalog-content\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.725147 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-utilities\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.725445 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct4nf\" (UniqueName: \"kubernetes.io/projected/8392e4b7-a4ea-4434-bf69-158870c1bfc6-kube-api-access-ct4nf\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.737777 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.827421 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct4nf\" (UniqueName: \"kubernetes.io/projected/8392e4b7-a4ea-4434-bf69-158870c1bfc6-kube-api-access-ct4nf\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.827534 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-catalog-content\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.827570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-utilities\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.828016 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-utilities\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.828114 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-catalog-content\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.845860 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct4nf\" (UniqueName: \"kubernetes.io/projected/8392e4b7-a4ea-4434-bf69-158870c1bfc6-kube-api-access-ct4nf\") pod \"redhat-marketplace-pd6pz\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:22 crc kubenswrapper[4758]: I1203 18:21:22.896083 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:23 crc kubenswrapper[4758]: I1203 18:21:23.114040 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:21:23 crc kubenswrapper[4758]: E1203 18:21:23.117233 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:21:23 crc kubenswrapper[4758]: I1203 18:21:23.246791 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6pz"] Dec 03 18:21:23 crc kubenswrapper[4758]: I1203 18:21:23.350573 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 03 18:21:23 crc kubenswrapper[4758]: W1203 18:21:23.367030 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30983465_84ac_482b_b11e_ceb9a2409ad3.slice/crio-ba10c95d6c28d2992a9504d653e6a65fd26a61a06e5f78fb170f9506ae4124cd WatchSource:0}: Error finding container ba10c95d6c28d2992a9504d653e6a65fd26a61a06e5f78fb170f9506ae4124cd: Status 404 returned error can't find the container with id ba10c95d6c28d2992a9504d653e6a65fd26a61a06e5f78fb170f9506ae4124cd Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.169291 4758 generic.go:334] "Generic (PLEG): container finished" podID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerID="2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971" exitCode=0 Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.169382 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6pz" event={"ID":"8392e4b7-a4ea-4434-bf69-158870c1bfc6","Type":"ContainerDied","Data":"2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971"} Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.169811 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6pz" event={"ID":"8392e4b7-a4ea-4434-bf69-158870c1bfc6","Type":"ContainerStarted","Data":"ecf157f8ee2711d727e5e6f1b0443b9a73c028d4df63075784ad442649b46910"} Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.173409 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.175748 4758 generic.go:334] "Generic (PLEG): container finished" podID="30983465-84ac-482b-b11e-ceb9a2409ad3" containerID="170a6f6514c8e4c51d323009bdfbde47cc633aceca2bc468502459ac0232341f" exitCode=0 Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.175806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"30983465-84ac-482b-b11e-ceb9a2409ad3","Type":"ContainerDied","Data":"170a6f6514c8e4c51d323009bdfbde47cc633aceca2bc468502459ac0232341f"} Dec 03 18:21:24 crc kubenswrapper[4758]: I1203 18:21:24.175840 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"30983465-84ac-482b-b11e-ceb9a2409ad3","Type":"ContainerStarted","Data":"ba10c95d6c28d2992a9504d653e6a65fd26a61a06e5f78fb170f9506ae4124cd"} Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.186214 4758 generic.go:334] "Generic (PLEG): container finished" podID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerID="326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33" exitCode=0 Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.186320 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6pz" event={"ID":"8392e4b7-a4ea-4434-bf69-158870c1bfc6","Type":"ContainerDied","Data":"326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33"} Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.524441 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.546351 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_30983465-84ac-482b-b11e-ceb9a2409ad3/mariadb-client-5-default/0.log" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.572770 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.588054 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.671438 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz2mh\" (UniqueName: \"kubernetes.io/projected/30983465-84ac-482b-b11e-ceb9a2409ad3-kube-api-access-nz2mh\") pod \"30983465-84ac-482b-b11e-ceb9a2409ad3\" (UID: \"30983465-84ac-482b-b11e-ceb9a2409ad3\") " Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.676491 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30983465-84ac-482b-b11e-ceb9a2409ad3-kube-api-access-nz2mh" (OuterVolumeSpecName: "kube-api-access-nz2mh") pod "30983465-84ac-482b-b11e-ceb9a2409ad3" (UID: "30983465-84ac-482b-b11e-ceb9a2409ad3"). InnerVolumeSpecName "kube-api-access-nz2mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.725911 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 03 18:21:25 crc kubenswrapper[4758]: E1203 18:21:25.726294 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30983465-84ac-482b-b11e-ceb9a2409ad3" containerName="mariadb-client-5-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.726319 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="30983465-84ac-482b-b11e-ceb9a2409ad3" containerName="mariadb-client-5-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.726512 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="30983465-84ac-482b-b11e-ceb9a2409ad3" containerName="mariadb-client-5-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.733447 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.733595 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.773711 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz2mh\" (UniqueName: \"kubernetes.io/projected/30983465-84ac-482b-b11e-ceb9a2409ad3-kube-api-access-nz2mh\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.875307 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24ps8\" (UniqueName: \"kubernetes.io/projected/d628e4ae-29bb-4e8f-b773-0e4a914244ae-kube-api-access-24ps8\") pod \"mariadb-client-6-default\" (UID: \"d628e4ae-29bb-4e8f-b773-0e4a914244ae\") " pod="openstack/mariadb-client-6-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.976884 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24ps8\" (UniqueName: \"kubernetes.io/projected/d628e4ae-29bb-4e8f-b773-0e4a914244ae-kube-api-access-24ps8\") pod \"mariadb-client-6-default\" (UID: \"d628e4ae-29bb-4e8f-b773-0e4a914244ae\") " pod="openstack/mariadb-client-6-default" Dec 03 18:21:25 crc kubenswrapper[4758]: I1203 18:21:25.993827 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24ps8\" (UniqueName: \"kubernetes.io/projected/d628e4ae-29bb-4e8f-b773-0e4a914244ae-kube-api-access-24ps8\") pod \"mariadb-client-6-default\" (UID: \"d628e4ae-29bb-4e8f-b773-0e4a914244ae\") " pod="openstack/mariadb-client-6-default" Dec 03 18:21:26 crc kubenswrapper[4758]: I1203 18:21:26.056320 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 03 18:21:26 crc kubenswrapper[4758]: I1203 18:21:26.233987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6pz" event={"ID":"8392e4b7-a4ea-4434-bf69-158870c1bfc6","Type":"ContainerStarted","Data":"0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2"} Dec 03 18:21:26 crc kubenswrapper[4758]: I1203 18:21:26.235672 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba10c95d6c28d2992a9504d653e6a65fd26a61a06e5f78fb170f9506ae4124cd" Dec 03 18:21:26 crc kubenswrapper[4758]: I1203 18:21:26.235754 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 03 18:21:26 crc kubenswrapper[4758]: I1203 18:21:26.260230 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pd6pz" podStartSLOduration=2.842899432 podStartE2EDuration="4.2602133s" podCreationTimestamp="2025-12-03 18:21:22 +0000 UTC" firstStartedPulling="2025-12-03 18:21:24.173017625 +0000 UTC m=+5139.374394506" lastFinishedPulling="2025-12-03 18:21:25.590331503 +0000 UTC m=+5140.791708374" observedRunningTime="2025-12-03 18:21:26.253515709 +0000 UTC m=+5141.454892570" watchObservedRunningTime="2025-12-03 18:21:26.2602133 +0000 UTC m=+5141.461590161" Dec 03 18:21:26 crc kubenswrapper[4758]: I1203 18:21:26.654705 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 03 18:21:27 crc kubenswrapper[4758]: I1203 18:21:27.124419 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30983465-84ac-482b-b11e-ceb9a2409ad3" path="/var/lib/kubelet/pods/30983465-84ac-482b-b11e-ceb9a2409ad3/volumes" Dec 03 18:21:27 crc kubenswrapper[4758]: I1203 18:21:27.243919 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"d628e4ae-29bb-4e8f-b773-0e4a914244ae","Type":"ContainerStarted","Data":"20c83482882c4646603895b967a564b19d61051fcfb9475c32608947171d8448"} Dec 03 18:21:27 crc kubenswrapper[4758]: I1203 18:21:27.243984 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"d628e4ae-29bb-4e8f-b773-0e4a914244ae","Type":"ContainerStarted","Data":"87f0ae0a0fd2581e88fdd315fb62d082a01a97939c27b9416521a2a65eccf7ff"} Dec 03 18:21:27 crc kubenswrapper[4758]: I1203 18:21:27.265548 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=2.265526894 podStartE2EDuration="2.265526894s" podCreationTimestamp="2025-12-03 18:21:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:21:27.257990619 +0000 UTC m=+5142.459367480" watchObservedRunningTime="2025-12-03 18:21:27.265526894 +0000 UTC m=+5142.466903765" Dec 03 18:21:28 crc kubenswrapper[4758]: I1203 18:21:28.256369 4758 generic.go:334] "Generic (PLEG): container finished" podID="d628e4ae-29bb-4e8f-b773-0e4a914244ae" containerID="20c83482882c4646603895b967a564b19d61051fcfb9475c32608947171d8448" exitCode=1 Dec 03 18:21:28 crc kubenswrapper[4758]: I1203 18:21:28.256446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"d628e4ae-29bb-4e8f-b773-0e4a914244ae","Type":"ContainerDied","Data":"20c83482882c4646603895b967a564b19d61051fcfb9475c32608947171d8448"} Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.717955 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.749512 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.759099 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.842201 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24ps8\" (UniqueName: \"kubernetes.io/projected/d628e4ae-29bb-4e8f-b773-0e4a914244ae-kube-api-access-24ps8\") pod \"d628e4ae-29bb-4e8f-b773-0e4a914244ae\" (UID: \"d628e4ae-29bb-4e8f-b773-0e4a914244ae\") " Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.847564 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d628e4ae-29bb-4e8f-b773-0e4a914244ae-kube-api-access-24ps8" (OuterVolumeSpecName: "kube-api-access-24ps8") pod "d628e4ae-29bb-4e8f-b773-0e4a914244ae" (UID: "d628e4ae-29bb-4e8f-b773-0e4a914244ae"). InnerVolumeSpecName "kube-api-access-24ps8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.918436 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 03 18:21:29 crc kubenswrapper[4758]: E1203 18:21:29.918852 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d628e4ae-29bb-4e8f-b773-0e4a914244ae" containerName="mariadb-client-6-default" Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.918875 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d628e4ae-29bb-4e8f-b773-0e4a914244ae" containerName="mariadb-client-6-default" Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.919083 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d628e4ae-29bb-4e8f-b773-0e4a914244ae" containerName="mariadb-client-6-default" Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.919756 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.928972 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 03 18:21:29 crc kubenswrapper[4758]: I1203 18:21:29.943463 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24ps8\" (UniqueName: \"kubernetes.io/projected/d628e4ae-29bb-4e8f-b773-0e4a914244ae-kube-api-access-24ps8\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.044589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ww4x\" (UniqueName: \"kubernetes.io/projected/24320205-3d63-45a0-8189-f7cc7b52352d-kube-api-access-4ww4x\") pod \"mariadb-client-7-default\" (UID: \"24320205-3d63-45a0-8189-f7cc7b52352d\") " pod="openstack/mariadb-client-7-default" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.146474 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ww4x\" (UniqueName: \"kubernetes.io/projected/24320205-3d63-45a0-8189-f7cc7b52352d-kube-api-access-4ww4x\") pod \"mariadb-client-7-default\" (UID: \"24320205-3d63-45a0-8189-f7cc7b52352d\") " pod="openstack/mariadb-client-7-default" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.162501 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ww4x\" (UniqueName: \"kubernetes.io/projected/24320205-3d63-45a0-8189-f7cc7b52352d-kube-api-access-4ww4x\") pod \"mariadb-client-7-default\" (UID: \"24320205-3d63-45a0-8189-f7cc7b52352d\") " pod="openstack/mariadb-client-7-default" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.236394 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.280711 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87f0ae0a0fd2581e88fdd315fb62d082a01a97939c27b9416521a2a65eccf7ff" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.280783 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 03 18:21:30 crc kubenswrapper[4758]: I1203 18:21:30.781482 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 03 18:21:30 crc kubenswrapper[4758]: W1203 18:21:30.788458 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24320205_3d63_45a0_8189_f7cc7b52352d.slice/crio-673832adf42b5b3552440cfb9e1d9663c8b207b659bf6ccf17f49c499b640a22 WatchSource:0}: Error finding container 673832adf42b5b3552440cfb9e1d9663c8b207b659bf6ccf17f49c499b640a22: Status 404 returned error can't find the container with id 673832adf42b5b3552440cfb9e1d9663c8b207b659bf6ccf17f49c499b640a22 Dec 03 18:21:31 crc kubenswrapper[4758]: I1203 18:21:31.124349 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d628e4ae-29bb-4e8f-b773-0e4a914244ae" path="/var/lib/kubelet/pods/d628e4ae-29bb-4e8f-b773-0e4a914244ae/volumes" Dec 03 18:21:31 crc kubenswrapper[4758]: I1203 18:21:31.292195 4758 generic.go:334] "Generic (PLEG): container finished" podID="24320205-3d63-45a0-8189-f7cc7b52352d" containerID="5b9a908934fa2673abce87a39335c304f7f795214c900c7ef84a5a2ec6925e8f" exitCode=0 Dec 03 18:21:31 crc kubenswrapper[4758]: I1203 18:21:31.292227 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"24320205-3d63-45a0-8189-f7cc7b52352d","Type":"ContainerDied","Data":"5b9a908934fa2673abce87a39335c304f7f795214c900c7ef84a5a2ec6925e8f"} Dec 03 18:21:31 crc kubenswrapper[4758]: I1203 18:21:31.292580 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"24320205-3d63-45a0-8189-f7cc7b52352d","Type":"ContainerStarted","Data":"673832adf42b5b3552440cfb9e1d9663c8b207b659bf6ccf17f49c499b640a22"} Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.210572 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.210642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.239828 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.658154 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.677917 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_24320205-3d63-45a0-8189-f7cc7b52352d/mariadb-client-7-default/0.log" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.712578 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.718184 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.791810 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ww4x\" (UniqueName: \"kubernetes.io/projected/24320205-3d63-45a0-8189-f7cc7b52352d-kube-api-access-4ww4x\") pod \"24320205-3d63-45a0-8189-f7cc7b52352d\" (UID: \"24320205-3d63-45a0-8189-f7cc7b52352d\") " Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.796485 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24320205-3d63-45a0-8189-f7cc7b52352d-kube-api-access-4ww4x" (OuterVolumeSpecName: "kube-api-access-4ww4x") pod "24320205-3d63-45a0-8189-f7cc7b52352d" (UID: "24320205-3d63-45a0-8189-f7cc7b52352d"). InnerVolumeSpecName "kube-api-access-4ww4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.894098 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ww4x\" (UniqueName: \"kubernetes.io/projected/24320205-3d63-45a0-8189-f7cc7b52352d-kube-api-access-4ww4x\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.899238 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 03 18:21:33 crc kubenswrapper[4758]: E1203 18:21:33.899647 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24320205-3d63-45a0-8189-f7cc7b52352d" containerName="mariadb-client-7-default" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.899667 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="24320205-3d63-45a0-8189-f7cc7b52352d" containerName="mariadb-client-7-default" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.899909 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="24320205-3d63-45a0-8189-f7cc7b52352d" containerName="mariadb-client-7-default" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.900579 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.910550 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 03 18:21:33 crc kubenswrapper[4758]: I1203 18:21:33.995097 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpj4b\" (UniqueName: \"kubernetes.io/projected/240dd60f-8adb-4c0f-9758-6534ad4e9292-kube-api-access-cpj4b\") pod \"mariadb-client-2\" (UID: \"240dd60f-8adb-4c0f-9758-6534ad4e9292\") " pod="openstack/mariadb-client-2" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.097096 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpj4b\" (UniqueName: \"kubernetes.io/projected/240dd60f-8adb-4c0f-9758-6534ad4e9292-kube-api-access-cpj4b\") pod \"mariadb-client-2\" (UID: \"240dd60f-8adb-4c0f-9758-6534ad4e9292\") " pod="openstack/mariadb-client-2" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.121750 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpj4b\" (UniqueName: \"kubernetes.io/projected/240dd60f-8adb-4c0f-9758-6534ad4e9292-kube-api-access-cpj4b\") pod \"mariadb-client-2\" (UID: \"240dd60f-8adb-4c0f-9758-6534ad4e9292\") " pod="openstack/mariadb-client-2" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.224653 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="673832adf42b5b3552440cfb9e1d9663c8b207b659bf6ccf17f49c499b640a22" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.224703 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.238164 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.281537 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.329051 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6pz"] Dec 03 18:21:34 crc kubenswrapper[4758]: I1203 18:21:34.804193 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 03 18:21:35 crc kubenswrapper[4758]: I1203 18:21:35.123595 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24320205-3d63-45a0-8189-f7cc7b52352d" path="/var/lib/kubelet/pods/24320205-3d63-45a0-8189-f7cc7b52352d/volumes" Dec 03 18:21:35 crc kubenswrapper[4758]: I1203 18:21:35.233249 4758 generic.go:334] "Generic (PLEG): container finished" podID="240dd60f-8adb-4c0f-9758-6534ad4e9292" containerID="ddff68893991da4ccd18a949f8960541b424c8d95e86711d26b41aae46bc94f6" exitCode=0 Dec 03 18:21:35 crc kubenswrapper[4758]: I1203 18:21:35.233346 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"240dd60f-8adb-4c0f-9758-6534ad4e9292","Type":"ContainerDied","Data":"ddff68893991da4ccd18a949f8960541b424c8d95e86711d26b41aae46bc94f6"} Dec 03 18:21:35 crc kubenswrapper[4758]: I1203 18:21:35.233390 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"240dd60f-8adb-4c0f-9758-6534ad4e9292","Type":"ContainerStarted","Data":"f61e70a7bf110ad804df0280ee6c30b6b72b26a04f7c3594c9a2f7a9e7458b2d"} Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.114386 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:21:36 crc kubenswrapper[4758]: E1203 18:21:36.115117 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.240973 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pd6pz" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="registry-server" containerID="cri-o://0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2" gracePeriod=2 Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.785176 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.791795 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.801412 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_240dd60f-8adb-4c0f-9758-6534ad4e9292/mariadb-client-2/0.log" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.828281 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.833973 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.942060 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct4nf\" (UniqueName: \"kubernetes.io/projected/8392e4b7-a4ea-4434-bf69-158870c1bfc6-kube-api-access-ct4nf\") pod \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.942269 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-utilities\") pod \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.942347 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpj4b\" (UniqueName: \"kubernetes.io/projected/240dd60f-8adb-4c0f-9758-6534ad4e9292-kube-api-access-cpj4b\") pod \"240dd60f-8adb-4c0f-9758-6534ad4e9292\" (UID: \"240dd60f-8adb-4c0f-9758-6534ad4e9292\") " Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.942379 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-catalog-content\") pod \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\" (UID: \"8392e4b7-a4ea-4434-bf69-158870c1bfc6\") " Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.943507 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-utilities" (OuterVolumeSpecName: "utilities") pod "8392e4b7-a4ea-4434-bf69-158870c1bfc6" (UID: "8392e4b7-a4ea-4434-bf69-158870c1bfc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.950446 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/240dd60f-8adb-4c0f-9758-6534ad4e9292-kube-api-access-cpj4b" (OuterVolumeSpecName: "kube-api-access-cpj4b") pod "240dd60f-8adb-4c0f-9758-6534ad4e9292" (UID: "240dd60f-8adb-4c0f-9758-6534ad4e9292"). InnerVolumeSpecName "kube-api-access-cpj4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.950575 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8392e4b7-a4ea-4434-bf69-158870c1bfc6-kube-api-access-ct4nf" (OuterVolumeSpecName: "kube-api-access-ct4nf") pod "8392e4b7-a4ea-4434-bf69-158870c1bfc6" (UID: "8392e4b7-a4ea-4434-bf69-158870c1bfc6"). InnerVolumeSpecName "kube-api-access-ct4nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:36 crc kubenswrapper[4758]: I1203 18:21:36.962176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8392e4b7-a4ea-4434-bf69-158870c1bfc6" (UID: "8392e4b7-a4ea-4434-bf69-158870c1bfc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.045492 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.045556 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpj4b\" (UniqueName: \"kubernetes.io/projected/240dd60f-8adb-4c0f-9758-6534ad4e9292-kube-api-access-cpj4b\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.045576 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8392e4b7-a4ea-4434-bf69-158870c1bfc6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.045594 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct4nf\" (UniqueName: \"kubernetes.io/projected/8392e4b7-a4ea-4434-bf69-158870c1bfc6-kube-api-access-ct4nf\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.129878 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="240dd60f-8adb-4c0f-9758-6534ad4e9292" path="/var/lib/kubelet/pods/240dd60f-8adb-4c0f-9758-6534ad4e9292/volumes" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.253591 4758 scope.go:117] "RemoveContainer" containerID="ddff68893991da4ccd18a949f8960541b424c8d95e86711d26b41aae46bc94f6" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.253647 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.256216 4758 generic.go:334] "Generic (PLEG): container finished" podID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerID="0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2" exitCode=0 Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.256255 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6pz" event={"ID":"8392e4b7-a4ea-4434-bf69-158870c1bfc6","Type":"ContainerDied","Data":"0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2"} Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.256281 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6pz" event={"ID":"8392e4b7-a4ea-4434-bf69-158870c1bfc6","Type":"ContainerDied","Data":"ecf157f8ee2711d727e5e6f1b0443b9a73c028d4df63075784ad442649b46910"} Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.256316 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6pz" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.278602 4758 scope.go:117] "RemoveContainer" containerID="0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.284957 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6pz"] Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.295638 4758 scope.go:117] "RemoveContainer" containerID="326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.296966 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6pz"] Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.312969 4758 scope.go:117] "RemoveContainer" containerID="2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.338484 4758 scope.go:117] "RemoveContainer" containerID="0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2" Dec 03 18:21:37 crc kubenswrapper[4758]: E1203 18:21:37.338986 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2\": container with ID starting with 0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2 not found: ID does not exist" containerID="0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.339026 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2"} err="failed to get container status \"0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2\": rpc error: code = NotFound desc = could not find container \"0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2\": container with ID starting with 0452dbe9096158a668a2aa95ff526c04c6ed68739deb6eeda1109d38a5c5aee2 not found: ID does not exist" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.339052 4758 scope.go:117] "RemoveContainer" containerID="326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33" Dec 03 18:21:37 crc kubenswrapper[4758]: E1203 18:21:37.339488 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33\": container with ID starting with 326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33 not found: ID does not exist" containerID="326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.339522 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33"} err="failed to get container status \"326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33\": rpc error: code = NotFound desc = could not find container \"326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33\": container with ID starting with 326ed8038dc3e9e0d3f45643bdfd649a5cf5ce44895d9887f7073c6667fd5e33 not found: ID does not exist" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.339541 4758 scope.go:117] "RemoveContainer" containerID="2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971" Dec 03 18:21:37 crc kubenswrapper[4758]: E1203 18:21:37.340851 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971\": container with ID starting with 2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971 not found: ID does not exist" containerID="2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971" Dec 03 18:21:37 crc kubenswrapper[4758]: I1203 18:21:37.340895 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971"} err="failed to get container status \"2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971\": rpc error: code = NotFound desc = could not find container \"2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971\": container with ID starting with 2747d613172c7b9c95e29cdb9e7ba7026b8a43c82c84ef0194fe9c2a60e51971 not found: ID does not exist" Dec 03 18:21:39 crc kubenswrapper[4758]: I1203 18:21:39.128716 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" path="/var/lib/kubelet/pods/8392e4b7-a4ea-4434-bf69-158870c1bfc6/volumes" Dec 03 18:21:50 crc kubenswrapper[4758]: I1203 18:21:50.113960 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:21:50 crc kubenswrapper[4758]: E1203 18:21:50.115855 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.384577 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4x7hk"] Dec 03 18:21:58 crc kubenswrapper[4758]: E1203 18:21:58.386825 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="extract-content" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.386958 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="extract-content" Dec 03 18:21:58 crc kubenswrapper[4758]: E1203 18:21:58.387017 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="extract-utilities" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.387250 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="extract-utilities" Dec 03 18:21:58 crc kubenswrapper[4758]: E1203 18:21:58.387324 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="240dd60f-8adb-4c0f-9758-6534ad4e9292" containerName="mariadb-client-2" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.387437 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="240dd60f-8adb-4c0f-9758-6534ad4e9292" containerName="mariadb-client-2" Dec 03 18:21:58 crc kubenswrapper[4758]: E1203 18:21:58.387830 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="registry-server" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.387883 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="registry-server" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.388200 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8392e4b7-a4ea-4434-bf69-158870c1bfc6" containerName="registry-server" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.388279 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="240dd60f-8adb-4c0f-9758-6534ad4e9292" containerName="mariadb-client-2" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.392970 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.397056 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4x7hk"] Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.570923 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-catalog-content\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.571291 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-utilities\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.571347 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjblp\" (UniqueName: \"kubernetes.io/projected/876b94c4-2087-4691-92da-40f23aa5f638-kube-api-access-hjblp\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.672367 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-utilities\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.672432 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjblp\" (UniqueName: \"kubernetes.io/projected/876b94c4-2087-4691-92da-40f23aa5f638-kube-api-access-hjblp\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.672516 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-catalog-content\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.673055 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-catalog-content\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.673341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-utilities\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.698643 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjblp\" (UniqueName: \"kubernetes.io/projected/876b94c4-2087-4691-92da-40f23aa5f638-kube-api-access-hjblp\") pod \"certified-operators-4x7hk\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:58 crc kubenswrapper[4758]: I1203 18:21:58.719496 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:21:59 crc kubenswrapper[4758]: I1203 18:21:59.186984 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4x7hk"] Dec 03 18:21:59 crc kubenswrapper[4758]: I1203 18:21:59.434086 4758 generic.go:334] "Generic (PLEG): container finished" podID="876b94c4-2087-4691-92da-40f23aa5f638" containerID="99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2" exitCode=0 Dec 03 18:21:59 crc kubenswrapper[4758]: I1203 18:21:59.434149 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4x7hk" event={"ID":"876b94c4-2087-4691-92da-40f23aa5f638","Type":"ContainerDied","Data":"99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2"} Dec 03 18:21:59 crc kubenswrapper[4758]: I1203 18:21:59.434272 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4x7hk" event={"ID":"876b94c4-2087-4691-92da-40f23aa5f638","Type":"ContainerStarted","Data":"6cb3dfb278334ef2acb66c77c609962ab8b99372520d0d071ccd4e4e353ed266"} Dec 03 18:22:00 crc kubenswrapper[4758]: I1203 18:22:00.448552 4758 generic.go:334] "Generic (PLEG): container finished" podID="876b94c4-2087-4691-92da-40f23aa5f638" containerID="af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698" exitCode=0 Dec 03 18:22:00 crc kubenswrapper[4758]: I1203 18:22:00.448725 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4x7hk" event={"ID":"876b94c4-2087-4691-92da-40f23aa5f638","Type":"ContainerDied","Data":"af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698"} Dec 03 18:22:01 crc kubenswrapper[4758]: I1203 18:22:01.462215 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4x7hk" event={"ID":"876b94c4-2087-4691-92da-40f23aa5f638","Type":"ContainerStarted","Data":"f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83"} Dec 03 18:22:01 crc kubenswrapper[4758]: I1203 18:22:01.485523 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4x7hk" podStartSLOduration=2.020049196 podStartE2EDuration="3.485493264s" podCreationTimestamp="2025-12-03 18:21:58 +0000 UTC" firstStartedPulling="2025-12-03 18:21:59.436947883 +0000 UTC m=+5174.638324764" lastFinishedPulling="2025-12-03 18:22:00.902391971 +0000 UTC m=+5176.103768832" observedRunningTime="2025-12-03 18:22:01.480395596 +0000 UTC m=+5176.681772467" watchObservedRunningTime="2025-12-03 18:22:01.485493264 +0000 UTC m=+5176.686870115" Dec 03 18:22:02 crc kubenswrapper[4758]: I1203 18:22:02.115248 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:22:02 crc kubenswrapper[4758]: E1203 18:22:02.115899 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:22:08 crc kubenswrapper[4758]: I1203 18:22:08.720508 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:22:08 crc kubenswrapper[4758]: I1203 18:22:08.721096 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:22:08 crc kubenswrapper[4758]: I1203 18:22:08.775559 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:22:09 crc kubenswrapper[4758]: I1203 18:22:09.583048 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:22:09 crc kubenswrapper[4758]: I1203 18:22:09.630849 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4x7hk"] Dec 03 18:22:10 crc kubenswrapper[4758]: I1203 18:22:10.718537 4758 scope.go:117] "RemoveContainer" containerID="02f2c53787ea82adbb557c54ba14b4e72c526fe6a9d1dd1ad2cff6b3f9b5e37d" Dec 03 18:22:11 crc kubenswrapper[4758]: I1203 18:22:11.554404 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4x7hk" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="registry-server" containerID="cri-o://f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83" gracePeriod=2 Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.506390 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.563733 4758 generic.go:334] "Generic (PLEG): container finished" podID="876b94c4-2087-4691-92da-40f23aa5f638" containerID="f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83" exitCode=0 Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.563775 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4x7hk" event={"ID":"876b94c4-2087-4691-92da-40f23aa5f638","Type":"ContainerDied","Data":"f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83"} Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.563807 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4x7hk" event={"ID":"876b94c4-2087-4691-92da-40f23aa5f638","Type":"ContainerDied","Data":"6cb3dfb278334ef2acb66c77c609962ab8b99372520d0d071ccd4e4e353ed266"} Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.563809 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4x7hk" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.563826 4758 scope.go:117] "RemoveContainer" containerID="f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.590593 4758 scope.go:117] "RemoveContainer" containerID="af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.604778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjblp\" (UniqueName: \"kubernetes.io/projected/876b94c4-2087-4691-92da-40f23aa5f638-kube-api-access-hjblp\") pod \"876b94c4-2087-4691-92da-40f23aa5f638\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.604837 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-utilities\") pod \"876b94c4-2087-4691-92da-40f23aa5f638\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.604860 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-catalog-content\") pod \"876b94c4-2087-4691-92da-40f23aa5f638\" (UID: \"876b94c4-2087-4691-92da-40f23aa5f638\") " Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.606988 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-utilities" (OuterVolumeSpecName: "utilities") pod "876b94c4-2087-4691-92da-40f23aa5f638" (UID: "876b94c4-2087-4691-92da-40f23aa5f638"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.612470 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/876b94c4-2087-4691-92da-40f23aa5f638-kube-api-access-hjblp" (OuterVolumeSpecName: "kube-api-access-hjblp") pod "876b94c4-2087-4691-92da-40f23aa5f638" (UID: "876b94c4-2087-4691-92da-40f23aa5f638"). InnerVolumeSpecName "kube-api-access-hjblp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.620810 4758 scope.go:117] "RemoveContainer" containerID="99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.652040 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "876b94c4-2087-4691-92da-40f23aa5f638" (UID: "876b94c4-2087-4691-92da-40f23aa5f638"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.685777 4758 scope.go:117] "RemoveContainer" containerID="f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83" Dec 03 18:22:12 crc kubenswrapper[4758]: E1203 18:22:12.686397 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83\": container with ID starting with f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83 not found: ID does not exist" containerID="f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.686426 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83"} err="failed to get container status \"f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83\": rpc error: code = NotFound desc = could not find container \"f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83\": container with ID starting with f12489c67530d77e7d6fa16234ddb2befa3a5189bf519b3d852fb6abf547ae83 not found: ID does not exist" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.686447 4758 scope.go:117] "RemoveContainer" containerID="af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698" Dec 03 18:22:12 crc kubenswrapper[4758]: E1203 18:22:12.686921 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698\": container with ID starting with af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698 not found: ID does not exist" containerID="af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.686960 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698"} err="failed to get container status \"af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698\": rpc error: code = NotFound desc = could not find container \"af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698\": container with ID starting with af366a64e49fcaebee49b285baa3bf1b2929bd5a66cf1ad17d92e56a0cb56698 not found: ID does not exist" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.686979 4758 scope.go:117] "RemoveContainer" containerID="99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2" Dec 03 18:22:12 crc kubenswrapper[4758]: E1203 18:22:12.687250 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2\": container with ID starting with 99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2 not found: ID does not exist" containerID="99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.687297 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2"} err="failed to get container status \"99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2\": rpc error: code = NotFound desc = could not find container \"99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2\": container with ID starting with 99f0d12753436cb9656deccd6fce3d9b84d6a8f57c7c32e98f66a77504a169a2 not found: ID does not exist" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.706599 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjblp\" (UniqueName: \"kubernetes.io/projected/876b94c4-2087-4691-92da-40f23aa5f638-kube-api-access-hjblp\") on node \"crc\" DevicePath \"\"" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.706634 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.706647 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/876b94c4-2087-4691-92da-40f23aa5f638-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.901163 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4x7hk"] Dec 03 18:22:12 crc kubenswrapper[4758]: I1203 18:22:12.906802 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4x7hk"] Dec 03 18:22:13 crc kubenswrapper[4758]: I1203 18:22:13.131634 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="876b94c4-2087-4691-92da-40f23aa5f638" path="/var/lib/kubelet/pods/876b94c4-2087-4691-92da-40f23aa5f638/volumes" Dec 03 18:22:14 crc kubenswrapper[4758]: I1203 18:22:14.114720 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:22:14 crc kubenswrapper[4758]: E1203 18:22:14.115123 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:22:28 crc kubenswrapper[4758]: I1203 18:22:28.114624 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:22:28 crc kubenswrapper[4758]: E1203 18:22:28.115566 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:22:42 crc kubenswrapper[4758]: I1203 18:22:42.115340 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:22:42 crc kubenswrapper[4758]: E1203 18:22:42.116869 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:22:55 crc kubenswrapper[4758]: I1203 18:22:55.122838 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:22:55 crc kubenswrapper[4758]: E1203 18:22:55.124110 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:23:10 crc kubenswrapper[4758]: I1203 18:23:10.115040 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:23:10 crc kubenswrapper[4758]: E1203 18:23:10.116266 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:23:25 crc kubenswrapper[4758]: I1203 18:23:25.122976 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:23:25 crc kubenswrapper[4758]: E1203 18:23:25.126416 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:23:38 crc kubenswrapper[4758]: I1203 18:23:38.114476 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:23:38 crc kubenswrapper[4758]: E1203 18:23:38.115155 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:23:49 crc kubenswrapper[4758]: I1203 18:23:49.114865 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:23:49 crc kubenswrapper[4758]: E1203 18:23:49.115865 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:24:00 crc kubenswrapper[4758]: I1203 18:24:00.114059 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:24:00 crc kubenswrapper[4758]: E1203 18:24:00.114831 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:24:13 crc kubenswrapper[4758]: I1203 18:24:13.114414 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:24:13 crc kubenswrapper[4758]: E1203 18:24:13.115310 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.581446 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bcndd"] Dec 03 18:24:21 crc kubenswrapper[4758]: E1203 18:24:21.583076 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="registry-server" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.583107 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="registry-server" Dec 03 18:24:21 crc kubenswrapper[4758]: E1203 18:24:21.583138 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="extract-utilities" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.583147 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="extract-utilities" Dec 03 18:24:21 crc kubenswrapper[4758]: E1203 18:24:21.583159 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="extract-content" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.583164 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="extract-content" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.583478 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="876b94c4-2087-4691-92da-40f23aa5f638" containerName="registry-server" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.584791 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.587139 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcndd"] Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.622285 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-utilities\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.622337 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-catalog-content\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.622410 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6v7l\" (UniqueName: \"kubernetes.io/projected/b9463a13-bf53-44fd-91fa-71f46de3d17c-kube-api-access-h6v7l\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.724274 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6v7l\" (UniqueName: \"kubernetes.io/projected/b9463a13-bf53-44fd-91fa-71f46de3d17c-kube-api-access-h6v7l\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.724359 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-utilities\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.724406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-catalog-content\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.725180 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-catalog-content\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.725230 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-utilities\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.743140 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6v7l\" (UniqueName: \"kubernetes.io/projected/b9463a13-bf53-44fd-91fa-71f46de3d17c-kube-api-access-h6v7l\") pod \"redhat-operators-bcndd\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:21 crc kubenswrapper[4758]: I1203 18:24:21.905422 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:22 crc kubenswrapper[4758]: I1203 18:24:22.340623 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcndd"] Dec 03 18:24:22 crc kubenswrapper[4758]: I1203 18:24:22.808755 4758 generic.go:334] "Generic (PLEG): container finished" podID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerID="8ed10f57c2d1cd77dbc10b0c5de074456e32c4ce9f49ef79db16b17ef2d9a9d6" exitCode=0 Dec 03 18:24:22 crc kubenswrapper[4758]: I1203 18:24:22.808867 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcndd" event={"ID":"b9463a13-bf53-44fd-91fa-71f46de3d17c","Type":"ContainerDied","Data":"8ed10f57c2d1cd77dbc10b0c5de074456e32c4ce9f49ef79db16b17ef2d9a9d6"} Dec 03 18:24:22 crc kubenswrapper[4758]: I1203 18:24:22.809070 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcndd" event={"ID":"b9463a13-bf53-44fd-91fa-71f46de3d17c","Type":"ContainerStarted","Data":"65d848fad2d85ceb95cd93b8288b790ec4c6ea15484d69a0729a8ccd9e48c3f6"} Dec 03 18:24:24 crc kubenswrapper[4758]: I1203 18:24:24.829366 4758 generic.go:334] "Generic (PLEG): container finished" podID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerID="cf4fb6312395a698dc8d5f42320ed59132a9fd38cdb4103719fed1d9b04b555e" exitCode=0 Dec 03 18:24:24 crc kubenswrapper[4758]: I1203 18:24:24.829446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcndd" event={"ID":"b9463a13-bf53-44fd-91fa-71f46de3d17c","Type":"ContainerDied","Data":"cf4fb6312395a698dc8d5f42320ed59132a9fd38cdb4103719fed1d9b04b555e"} Dec 03 18:24:25 crc kubenswrapper[4758]: I1203 18:24:25.841509 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcndd" event={"ID":"b9463a13-bf53-44fd-91fa-71f46de3d17c","Type":"ContainerStarted","Data":"495bfef64a18f72f7659681862fd8484a23faff0862156ee140b4ba8744333b8"} Dec 03 18:24:25 crc kubenswrapper[4758]: I1203 18:24:25.872882 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bcndd" podStartSLOduration=2.389589299 podStartE2EDuration="4.872860128s" podCreationTimestamp="2025-12-03 18:24:21 +0000 UTC" firstStartedPulling="2025-12-03 18:24:22.81034448 +0000 UTC m=+5318.011721341" lastFinishedPulling="2025-12-03 18:24:25.293615299 +0000 UTC m=+5320.494992170" observedRunningTime="2025-12-03 18:24:25.866420044 +0000 UTC m=+5321.067796915" watchObservedRunningTime="2025-12-03 18:24:25.872860128 +0000 UTC m=+5321.074236989" Dec 03 18:24:28 crc kubenswrapper[4758]: I1203 18:24:28.113937 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:24:28 crc kubenswrapper[4758]: E1203 18:24:28.114389 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:24:31 crc kubenswrapper[4758]: I1203 18:24:31.906583 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:31 crc kubenswrapper[4758]: I1203 18:24:31.907394 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:32 crc kubenswrapper[4758]: I1203 18:24:32.989962 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bcndd" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="registry-server" probeResult="failure" output=< Dec 03 18:24:32 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 18:24:32 crc kubenswrapper[4758]: > Dec 03 18:24:39 crc kubenswrapper[4758]: I1203 18:24:39.115035 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:24:39 crc kubenswrapper[4758]: E1203 18:24:39.115597 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:24:41 crc kubenswrapper[4758]: I1203 18:24:41.975810 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:42 crc kubenswrapper[4758]: I1203 18:24:42.040104 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:42 crc kubenswrapper[4758]: I1203 18:24:42.223188 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcndd"] Dec 03 18:24:44 crc kubenswrapper[4758]: I1203 18:24:44.004880 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bcndd" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="registry-server" containerID="cri-o://495bfef64a18f72f7659681862fd8484a23faff0862156ee140b4ba8744333b8" gracePeriod=2 Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.046242 4758 generic.go:334] "Generic (PLEG): container finished" podID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerID="495bfef64a18f72f7659681862fd8484a23faff0862156ee140b4ba8744333b8" exitCode=0 Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.046466 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcndd" event={"ID":"b9463a13-bf53-44fd-91fa-71f46de3d17c","Type":"ContainerDied","Data":"495bfef64a18f72f7659681862fd8484a23faff0862156ee140b4ba8744333b8"} Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.345346 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.481185 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-utilities\") pod \"b9463a13-bf53-44fd-91fa-71f46de3d17c\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.481291 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6v7l\" (UniqueName: \"kubernetes.io/projected/b9463a13-bf53-44fd-91fa-71f46de3d17c-kube-api-access-h6v7l\") pod \"b9463a13-bf53-44fd-91fa-71f46de3d17c\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.481322 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-catalog-content\") pod \"b9463a13-bf53-44fd-91fa-71f46de3d17c\" (UID: \"b9463a13-bf53-44fd-91fa-71f46de3d17c\") " Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.483119 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-utilities" (OuterVolumeSpecName: "utilities") pod "b9463a13-bf53-44fd-91fa-71f46de3d17c" (UID: "b9463a13-bf53-44fd-91fa-71f46de3d17c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.488989 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9463a13-bf53-44fd-91fa-71f46de3d17c-kube-api-access-h6v7l" (OuterVolumeSpecName: "kube-api-access-h6v7l") pod "b9463a13-bf53-44fd-91fa-71f46de3d17c" (UID: "b9463a13-bf53-44fd-91fa-71f46de3d17c"). InnerVolumeSpecName "kube-api-access-h6v7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.582643 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6v7l\" (UniqueName: \"kubernetes.io/projected/b9463a13-bf53-44fd-91fa-71f46de3d17c-kube-api-access-h6v7l\") on node \"crc\" DevicePath \"\"" Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.582734 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.594278 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9463a13-bf53-44fd-91fa-71f46de3d17c" (UID: "b9463a13-bf53-44fd-91fa-71f46de3d17c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:24:46 crc kubenswrapper[4758]: I1203 18:24:46.683655 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9463a13-bf53-44fd-91fa-71f46de3d17c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.058911 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcndd" event={"ID":"b9463a13-bf53-44fd-91fa-71f46de3d17c","Type":"ContainerDied","Data":"65d848fad2d85ceb95cd93b8288b790ec4c6ea15484d69a0729a8ccd9e48c3f6"} Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.058984 4758 scope.go:117] "RemoveContainer" containerID="495bfef64a18f72f7659681862fd8484a23faff0862156ee140b4ba8744333b8" Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.059256 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcndd" Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.085620 4758 scope.go:117] "RemoveContainer" containerID="cf4fb6312395a698dc8d5f42320ed59132a9fd38cdb4103719fed1d9b04b555e" Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.112352 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcndd"] Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.133032 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bcndd"] Dec 03 18:24:47 crc kubenswrapper[4758]: I1203 18:24:47.140347 4758 scope.go:117] "RemoveContainer" containerID="8ed10f57c2d1cd77dbc10b0c5de074456e32c4ce9f49ef79db16b17ef2d9a9d6" Dec 03 18:24:49 crc kubenswrapper[4758]: I1203 18:24:49.145471 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" path="/var/lib/kubelet/pods/b9463a13-bf53-44fd-91fa-71f46de3d17c/volumes" Dec 03 18:24:54 crc kubenswrapper[4758]: I1203 18:24:54.113836 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:24:54 crc kubenswrapper[4758]: E1203 18:24:54.114376 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:25:06 crc kubenswrapper[4758]: I1203 18:25:06.115140 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:25:06 crc kubenswrapper[4758]: E1203 18:25:06.116347 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:25:18 crc kubenswrapper[4758]: I1203 18:25:18.114601 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:25:18 crc kubenswrapper[4758]: E1203 18:25:18.115539 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:25:32 crc kubenswrapper[4758]: I1203 18:25:32.115051 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:25:32 crc kubenswrapper[4758]: E1203 18:25:32.116141 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:25:46 crc kubenswrapper[4758]: I1203 18:25:46.114537 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:25:46 crc kubenswrapper[4758]: I1203 18:25:46.566108 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"ee02f3819ed4276671e01c5782910e87da9be6dd0555ff89c3f0a62d67c490da"} Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.154890 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 03 18:25:47 crc kubenswrapper[4758]: E1203 18:25:47.155185 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="registry-server" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.155197 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="registry-server" Dec 03 18:25:47 crc kubenswrapper[4758]: E1203 18:25:47.155225 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="extract-content" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.155231 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="extract-content" Dec 03 18:25:47 crc kubenswrapper[4758]: E1203 18:25:47.155245 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="extract-utilities" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.155253 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="extract-utilities" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.155377 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9463a13-bf53-44fd-91fa-71f46de3d17c" containerName="registry-server" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.155938 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.157593 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrxtm" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.172299 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.306240 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvq6h\" (UniqueName: \"kubernetes.io/projected/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c-kube-api-access-lvq6h\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.306305 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e00c4833-294a-47f9-9733-cdb516be49b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.408879 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvq6h\" (UniqueName: \"kubernetes.io/projected/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c-kube-api-access-lvq6h\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.409034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e00c4833-294a-47f9-9733-cdb516be49b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.413957 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.414010 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e00c4833-294a-47f9-9733-cdb516be49b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3e49afd83347ecdc594c9bb11c657ff639036d26747d10e3721ea955d339c271/globalmount\"" pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.428904 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvq6h\" (UniqueName: \"kubernetes.io/projected/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c-kube-api-access-lvq6h\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.443876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e00c4833-294a-47f9-9733-cdb516be49b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") pod \"mariadb-copy-data\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " pod="openstack/mariadb-copy-data" Dec 03 18:25:47 crc kubenswrapper[4758]: I1203 18:25:47.472716 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 03 18:25:48 crc kubenswrapper[4758]: I1203 18:25:48.014356 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 03 18:25:48 crc kubenswrapper[4758]: I1203 18:25:48.583486 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c","Type":"ContainerStarted","Data":"befb8fff4e1f597d0652fa0d78402245fbbbd8e53e64add401f1a71f8515d9dd"} Dec 03 18:25:48 crc kubenswrapper[4758]: I1203 18:25:48.583919 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c","Type":"ContainerStarted","Data":"f263d2e57e84a74476dae8318e6d6ec2941922f0f757e4ecd7117187ee23bb3e"} Dec 03 18:25:48 crc kubenswrapper[4758]: I1203 18:25:48.601707 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.6016639660000003 podStartE2EDuration="2.601663966s" podCreationTimestamp="2025-12-03 18:25:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:25:48.595912981 +0000 UTC m=+5403.797289842" watchObservedRunningTime="2025-12-03 18:25:48.601663966 +0000 UTC m=+5403.803040837" Dec 03 18:25:51 crc kubenswrapper[4758]: I1203 18:25:51.696556 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:51 crc kubenswrapper[4758]: I1203 18:25:51.698336 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:51 crc kubenswrapper[4758]: I1203 18:25:51.710953 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:51 crc kubenswrapper[4758]: I1203 18:25:51.882305 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ls58\" (UniqueName: \"kubernetes.io/projected/5868b5f5-466a-4b6b-acd4-fb7497005cf1-kube-api-access-8ls58\") pod \"mariadb-client\" (UID: \"5868b5f5-466a-4b6b-acd4-fb7497005cf1\") " pod="openstack/mariadb-client" Dec 03 18:25:51 crc kubenswrapper[4758]: I1203 18:25:51.984873 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ls58\" (UniqueName: \"kubernetes.io/projected/5868b5f5-466a-4b6b-acd4-fb7497005cf1-kube-api-access-8ls58\") pod \"mariadb-client\" (UID: \"5868b5f5-466a-4b6b-acd4-fb7497005cf1\") " pod="openstack/mariadb-client" Dec 03 18:25:52 crc kubenswrapper[4758]: I1203 18:25:52.005652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ls58\" (UniqueName: \"kubernetes.io/projected/5868b5f5-466a-4b6b-acd4-fb7497005cf1-kube-api-access-8ls58\") pod \"mariadb-client\" (UID: \"5868b5f5-466a-4b6b-acd4-fb7497005cf1\") " pod="openstack/mariadb-client" Dec 03 18:25:52 crc kubenswrapper[4758]: I1203 18:25:52.030105 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:52 crc kubenswrapper[4758]: I1203 18:25:52.241718 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:52 crc kubenswrapper[4758]: W1203 18:25:52.250884 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5868b5f5_466a_4b6b_acd4_fb7497005cf1.slice/crio-51246eae4cb6146f408d37036b4562278bd570015d2f12286a56f5b625448f84 WatchSource:0}: Error finding container 51246eae4cb6146f408d37036b4562278bd570015d2f12286a56f5b625448f84: Status 404 returned error can't find the container with id 51246eae4cb6146f408d37036b4562278bd570015d2f12286a56f5b625448f84 Dec 03 18:25:52 crc kubenswrapper[4758]: I1203 18:25:52.616303 4758 generic.go:334] "Generic (PLEG): container finished" podID="5868b5f5-466a-4b6b-acd4-fb7497005cf1" containerID="175644fa55952194731916bbad3b39a444c476323d83d2b88c2f7cd7ee4c525a" exitCode=0 Dec 03 18:25:52 crc kubenswrapper[4758]: I1203 18:25:52.616356 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"5868b5f5-466a-4b6b-acd4-fb7497005cf1","Type":"ContainerDied","Data":"175644fa55952194731916bbad3b39a444c476323d83d2b88c2f7cd7ee4c525a"} Dec 03 18:25:52 crc kubenswrapper[4758]: I1203 18:25:52.616407 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"5868b5f5-466a-4b6b-acd4-fb7497005cf1","Type":"ContainerStarted","Data":"51246eae4cb6146f408d37036b4562278bd570015d2f12286a56f5b625448f84"} Dec 03 18:25:53 crc kubenswrapper[4758]: I1203 18:25:53.962901 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:53 crc kubenswrapper[4758]: I1203 18:25:53.986018 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_5868b5f5-466a-4b6b-acd4-fb7497005cf1/mariadb-client/0.log" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.014513 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.016226 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ls58\" (UniqueName: \"kubernetes.io/projected/5868b5f5-466a-4b6b-acd4-fb7497005cf1-kube-api-access-8ls58\") pod \"5868b5f5-466a-4b6b-acd4-fb7497005cf1\" (UID: \"5868b5f5-466a-4b6b-acd4-fb7497005cf1\") " Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.020836 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.021832 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5868b5f5-466a-4b6b-acd4-fb7497005cf1-kube-api-access-8ls58" (OuterVolumeSpecName: "kube-api-access-8ls58") pod "5868b5f5-466a-4b6b-acd4-fb7497005cf1" (UID: "5868b5f5-466a-4b6b-acd4-fb7497005cf1"). InnerVolumeSpecName "kube-api-access-8ls58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.118582 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ls58\" (UniqueName: \"kubernetes.io/projected/5868b5f5-466a-4b6b-acd4-fb7497005cf1-kube-api-access-8ls58\") on node \"crc\" DevicePath \"\"" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.196718 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:54 crc kubenswrapper[4758]: E1203 18:25:54.197154 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5868b5f5-466a-4b6b-acd4-fb7497005cf1" containerName="mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.197172 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5868b5f5-466a-4b6b-acd4-fb7497005cf1" containerName="mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.197325 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5868b5f5-466a-4b6b-acd4-fb7497005cf1" containerName="mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.197931 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.203374 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.323558 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmb87\" (UniqueName: \"kubernetes.io/projected/ed761b49-17ff-4087-8050-ec397a6480c9-kube-api-access-kmb87\") pod \"mariadb-client\" (UID: \"ed761b49-17ff-4087-8050-ec397a6480c9\") " pod="openstack/mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.424771 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmb87\" (UniqueName: \"kubernetes.io/projected/ed761b49-17ff-4087-8050-ec397a6480c9-kube-api-access-kmb87\") pod \"mariadb-client\" (UID: \"ed761b49-17ff-4087-8050-ec397a6480c9\") " pod="openstack/mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.443969 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmb87\" (UniqueName: \"kubernetes.io/projected/ed761b49-17ff-4087-8050-ec397a6480c9-kube-api-access-kmb87\") pod \"mariadb-client\" (UID: \"ed761b49-17ff-4087-8050-ec397a6480c9\") " pod="openstack/mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.525023 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.636385 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51246eae4cb6146f408d37036b4562278bd570015d2f12286a56f5b625448f84" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.636465 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.664804 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="5868b5f5-466a-4b6b-acd4-fb7497005cf1" podUID="ed761b49-17ff-4087-8050-ec397a6480c9" Dec 03 18:25:54 crc kubenswrapper[4758]: I1203 18:25:54.967021 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:54 crc kubenswrapper[4758]: W1203 18:25:54.975104 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded761b49_17ff_4087_8050_ec397a6480c9.slice/crio-a2dd6d1bf0cf4bc4cd84a48b6fe81ad775e7f531df530addd825f77356733121 WatchSource:0}: Error finding container a2dd6d1bf0cf4bc4cd84a48b6fe81ad775e7f531df530addd825f77356733121: Status 404 returned error can't find the container with id a2dd6d1bf0cf4bc4cd84a48b6fe81ad775e7f531df530addd825f77356733121 Dec 03 18:25:55 crc kubenswrapper[4758]: I1203 18:25:55.123749 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5868b5f5-466a-4b6b-acd4-fb7497005cf1" path="/var/lib/kubelet/pods/5868b5f5-466a-4b6b-acd4-fb7497005cf1/volumes" Dec 03 18:25:55 crc kubenswrapper[4758]: I1203 18:25:55.644914 4758 generic.go:334] "Generic (PLEG): container finished" podID="ed761b49-17ff-4087-8050-ec397a6480c9" containerID="dbcdfca6efeb07ea11031a14a774b324e228674c9d43fc9e09c1ab64f8dfde6b" exitCode=0 Dec 03 18:25:55 crc kubenswrapper[4758]: I1203 18:25:55.645003 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"ed761b49-17ff-4087-8050-ec397a6480c9","Type":"ContainerDied","Data":"dbcdfca6efeb07ea11031a14a774b324e228674c9d43fc9e09c1ab64f8dfde6b"} Dec 03 18:25:55 crc kubenswrapper[4758]: I1203 18:25:55.645285 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"ed761b49-17ff-4087-8050-ec397a6480c9","Type":"ContainerStarted","Data":"a2dd6d1bf0cf4bc4cd84a48b6fe81ad775e7f531df530addd825f77356733121"} Dec 03 18:25:56 crc kubenswrapper[4758]: I1203 18:25:56.946302 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:25:56 crc kubenswrapper[4758]: I1203 18:25:56.967152 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_ed761b49-17ff-4087-8050-ec397a6480c9/mariadb-client/0.log" Dec 03 18:25:56 crc kubenswrapper[4758]: I1203 18:25:56.994817 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.002030 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.065363 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmb87\" (UniqueName: \"kubernetes.io/projected/ed761b49-17ff-4087-8050-ec397a6480c9-kube-api-access-kmb87\") pod \"ed761b49-17ff-4087-8050-ec397a6480c9\" (UID: \"ed761b49-17ff-4087-8050-ec397a6480c9\") " Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.071025 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed761b49-17ff-4087-8050-ec397a6480c9-kube-api-access-kmb87" (OuterVolumeSpecName: "kube-api-access-kmb87") pod "ed761b49-17ff-4087-8050-ec397a6480c9" (UID: "ed761b49-17ff-4087-8050-ec397a6480c9"). InnerVolumeSpecName "kube-api-access-kmb87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.125396 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed761b49-17ff-4087-8050-ec397a6480c9" path="/var/lib/kubelet/pods/ed761b49-17ff-4087-8050-ec397a6480c9/volumes" Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.166911 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmb87\" (UniqueName: \"kubernetes.io/projected/ed761b49-17ff-4087-8050-ec397a6480c9-kube-api-access-kmb87\") on node \"crc\" DevicePath \"\"" Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.660253 4758 scope.go:117] "RemoveContainer" containerID="dbcdfca6efeb07ea11031a14a774b324e228674c9d43fc9e09c1ab64f8dfde6b" Dec 03 18:25:57 crc kubenswrapper[4758]: I1203 18:25:57.660288 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.382065 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2lzkw"] Dec 03 18:26:02 crc kubenswrapper[4758]: E1203 18:26:02.382990 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed761b49-17ff-4087-8050-ec397a6480c9" containerName="mariadb-client" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.383009 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed761b49-17ff-4087-8050-ec397a6480c9" containerName="mariadb-client" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.383182 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed761b49-17ff-4087-8050-ec397a6480c9" containerName="mariadb-client" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.386266 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.403598 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lzkw"] Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.454969 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7vlc\" (UniqueName: \"kubernetes.io/projected/721a6329-c1b0-4f18-8218-927e8f1d2a95-kube-api-access-t7vlc\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.455070 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-catalog-content\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.455138 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-utilities\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.555948 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7vlc\" (UniqueName: \"kubernetes.io/projected/721a6329-c1b0-4f18-8218-927e8f1d2a95-kube-api-access-t7vlc\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.556039 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-catalog-content\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.556084 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-utilities\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.556922 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-utilities\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.557151 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-catalog-content\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.575768 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7vlc\" (UniqueName: \"kubernetes.io/projected/721a6329-c1b0-4f18-8218-927e8f1d2a95-kube-api-access-t7vlc\") pod \"community-operators-2lzkw\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:02 crc kubenswrapper[4758]: I1203 18:26:02.719190 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:03 crc kubenswrapper[4758]: I1203 18:26:03.210161 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lzkw"] Dec 03 18:26:03 crc kubenswrapper[4758]: W1203 18:26:03.210892 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721a6329_c1b0_4f18_8218_927e8f1d2a95.slice/crio-dfa777a4a17e746be3a9f6e8b650210598a3df8246c8694c21843ad2308b753c WatchSource:0}: Error finding container dfa777a4a17e746be3a9f6e8b650210598a3df8246c8694c21843ad2308b753c: Status 404 returned error can't find the container with id dfa777a4a17e746be3a9f6e8b650210598a3df8246c8694c21843ad2308b753c Dec 03 18:26:03 crc kubenswrapper[4758]: I1203 18:26:03.710418 4758 generic.go:334] "Generic (PLEG): container finished" podID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerID="0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2" exitCode=0 Dec 03 18:26:03 crc kubenswrapper[4758]: I1203 18:26:03.710534 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerDied","Data":"0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2"} Dec 03 18:26:03 crc kubenswrapper[4758]: I1203 18:26:03.712005 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerStarted","Data":"dfa777a4a17e746be3a9f6e8b650210598a3df8246c8694c21843ad2308b753c"} Dec 03 18:26:05 crc kubenswrapper[4758]: I1203 18:26:05.733385 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerStarted","Data":"d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599"} Dec 03 18:26:06 crc kubenswrapper[4758]: I1203 18:26:06.749803 4758 generic.go:334] "Generic (PLEG): container finished" podID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerID="d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599" exitCode=0 Dec 03 18:26:06 crc kubenswrapper[4758]: I1203 18:26:06.749946 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerDied","Data":"d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599"} Dec 03 18:26:08 crc kubenswrapper[4758]: I1203 18:26:08.769840 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerStarted","Data":"c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089"} Dec 03 18:26:08 crc kubenswrapper[4758]: I1203 18:26:08.789811 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2lzkw" podStartSLOduration=2.752802488 podStartE2EDuration="6.789791296s" podCreationTimestamp="2025-12-03 18:26:02 +0000 UTC" firstStartedPulling="2025-12-03 18:26:03.713169477 +0000 UTC m=+5418.914546358" lastFinishedPulling="2025-12-03 18:26:07.750158305 +0000 UTC m=+5422.951535166" observedRunningTime="2025-12-03 18:26:08.78549765 +0000 UTC m=+5423.986874511" watchObservedRunningTime="2025-12-03 18:26:08.789791296 +0000 UTC m=+5423.991168157" Dec 03 18:26:12 crc kubenswrapper[4758]: I1203 18:26:12.720073 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:12 crc kubenswrapper[4758]: I1203 18:26:12.720734 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:12 crc kubenswrapper[4758]: I1203 18:26:12.784642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:12 crc kubenswrapper[4758]: I1203 18:26:12.851882 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:13 crc kubenswrapper[4758]: I1203 18:26:13.020981 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lzkw"] Dec 03 18:26:14 crc kubenswrapper[4758]: I1203 18:26:14.826880 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2lzkw" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="registry-server" containerID="cri-o://c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089" gracePeriod=2 Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.287733 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.468360 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-catalog-content\") pod \"721a6329-c1b0-4f18-8218-927e8f1d2a95\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.468413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7vlc\" (UniqueName: \"kubernetes.io/projected/721a6329-c1b0-4f18-8218-927e8f1d2a95-kube-api-access-t7vlc\") pod \"721a6329-c1b0-4f18-8218-927e8f1d2a95\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.468443 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-utilities\") pod \"721a6329-c1b0-4f18-8218-927e8f1d2a95\" (UID: \"721a6329-c1b0-4f18-8218-927e8f1d2a95\") " Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.469549 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-utilities" (OuterVolumeSpecName: "utilities") pod "721a6329-c1b0-4f18-8218-927e8f1d2a95" (UID: "721a6329-c1b0-4f18-8218-927e8f1d2a95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.476873 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/721a6329-c1b0-4f18-8218-927e8f1d2a95-kube-api-access-t7vlc" (OuterVolumeSpecName: "kube-api-access-t7vlc") pod "721a6329-c1b0-4f18-8218-927e8f1d2a95" (UID: "721a6329-c1b0-4f18-8218-927e8f1d2a95"). InnerVolumeSpecName "kube-api-access-t7vlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.532997 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "721a6329-c1b0-4f18-8218-927e8f1d2a95" (UID: "721a6329-c1b0-4f18-8218-927e8f1d2a95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.570628 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.570967 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7vlc\" (UniqueName: \"kubernetes.io/projected/721a6329-c1b0-4f18-8218-927e8f1d2a95-kube-api-access-t7vlc\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.571115 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/721a6329-c1b0-4f18-8218-927e8f1d2a95-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.836874 4758 generic.go:334] "Generic (PLEG): container finished" podID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerID="c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089" exitCode=0 Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.836883 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerDied","Data":"c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089"} Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.838261 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lzkw" event={"ID":"721a6329-c1b0-4f18-8218-927e8f1d2a95","Type":"ContainerDied","Data":"dfa777a4a17e746be3a9f6e8b650210598a3df8246c8694c21843ad2308b753c"} Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.838339 4758 scope.go:117] "RemoveContainer" containerID="c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.836896 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lzkw" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.871742 4758 scope.go:117] "RemoveContainer" containerID="d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.885066 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lzkw"] Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.892355 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2lzkw"] Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.907364 4758 scope.go:117] "RemoveContainer" containerID="0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.933276 4758 scope.go:117] "RemoveContainer" containerID="c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089" Dec 03 18:26:15 crc kubenswrapper[4758]: E1203 18:26:15.935329 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089\": container with ID starting with c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089 not found: ID does not exist" containerID="c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.935488 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089"} err="failed to get container status \"c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089\": rpc error: code = NotFound desc = could not find container \"c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089\": container with ID starting with c4a0aac61c04c41dc400c75de49bca31ea5654c3371efaef1605bb3d2b627089 not found: ID does not exist" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.935613 4758 scope.go:117] "RemoveContainer" containerID="d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599" Dec 03 18:26:15 crc kubenswrapper[4758]: E1203 18:26:15.936213 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599\": container with ID starting with d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599 not found: ID does not exist" containerID="d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.936340 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599"} err="failed to get container status \"d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599\": rpc error: code = NotFound desc = could not find container \"d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599\": container with ID starting with d234ccc70abfa5bb3ee02feef747535acf8188c47dd16e9de291ac79bd582599 not found: ID does not exist" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.936437 4758 scope.go:117] "RemoveContainer" containerID="0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2" Dec 03 18:26:15 crc kubenswrapper[4758]: E1203 18:26:15.936956 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2\": container with ID starting with 0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2 not found: ID does not exist" containerID="0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2" Dec 03 18:26:15 crc kubenswrapper[4758]: I1203 18:26:15.937199 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2"} err="failed to get container status \"0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2\": rpc error: code = NotFound desc = could not find container \"0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2\": container with ID starting with 0a526daca5c6e2c00419145886edbd4f4f18855625a5b30b5793f0dcee8ae7e2 not found: ID does not exist" Dec 03 18:26:17 crc kubenswrapper[4758]: I1203 18:26:17.123447 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" path="/var/lib/kubelet/pods/721a6329-c1b0-4f18-8218-927e8f1d2a95/volumes" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.587462 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 18:26:46 crc kubenswrapper[4758]: E1203 18:26:46.588584 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="extract-content" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.588611 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="extract-content" Dec 03 18:26:46 crc kubenswrapper[4758]: E1203 18:26:46.588634 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="registry-server" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.588646 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="registry-server" Dec 03 18:26:46 crc kubenswrapper[4758]: E1203 18:26:46.588709 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="extract-utilities" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.588722 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="extract-utilities" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.589121 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="721a6329-c1b0-4f18-8218-927e8f1d2a95" containerName="registry-server" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.590407 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.593829 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.593966 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-gh8r9" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.595320 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.601288 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.608759 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.610847 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.619779 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.622646 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.630189 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.642727 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.695221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6812355e-c51f-499b-bfe0-3d78d40e86c7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.695321 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv2bz\" (UniqueName: \"kubernetes.io/projected/6812355e-c51f-499b-bfe0-3d78d40e86c7-kube-api-access-xv2bz\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.695400 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6812355e-c51f-499b-bfe0-3d78d40e86c7-config\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.695455 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.695512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6812355e-c51f-499b-bfe0-3d78d40e86c7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.695545 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6812355e-c51f-499b-bfe0-3d78d40e86c7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.797299 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6812355e-c51f-499b-bfe0-3d78d40e86c7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.797600 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.797726 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6812355e-c51f-499b-bfe0-3d78d40e86c7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.797906 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d6fb8c-07d0-41f4-b027-76d126425a78-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.798089 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-config\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.798242 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6812355e-c51f-499b-bfe0-3d78d40e86c7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.798417 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7d6fb8c-07d0-41f4-b027-76d126425a78-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.798583 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zndcn\" (UniqueName: \"kubernetes.io/projected/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-kube-api-access-zndcn\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.798757 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799088 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv2bz\" (UniqueName: \"kubernetes.io/projected/6812355e-c51f-499b-bfe0-3d78d40e86c7-kube-api-access-xv2bz\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799170 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799252 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799286 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9clgw\" (UniqueName: \"kubernetes.io/projected/a7d6fb8c-07d0-41f4-b027-76d126425a78-kube-api-access-9clgw\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799313 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d6fb8c-07d0-41f4-b027-76d126425a78-config\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799357 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6812355e-c51f-499b-bfe0-3d78d40e86c7-config\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799452 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799507 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d6fb8c-07d0-41f4-b027-76d126425a78-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.799553 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.800726 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6812355e-c51f-499b-bfe0-3d78d40e86c7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.801073 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6812355e-c51f-499b-bfe0-3d78d40e86c7-config\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.801094 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6812355e-c51f-499b-bfe0-3d78d40e86c7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.803856 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.803986 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a8f4e811c9b27013ebeebb5b855090db2227bce85acee470042391a7cb3c2572/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.825161 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6812355e-c51f-499b-bfe0-3d78d40e86c7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.825322 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.835384 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.839585 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.840151 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.840153 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-4b44w" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.845439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv2bz\" (UniqueName: \"kubernetes.io/projected/6812355e-c51f-499b-bfe0-3d78d40e86c7-kube-api-access-xv2bz\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.850301 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.861489 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.864615 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.866084 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4e915e2-4f68-44ad-85d7-159f267c1cc5\") pod \"ovsdbserver-sb-0\" (UID: \"6812355e-c51f-499b-bfe0-3d78d40e86c7\") " pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.868670 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.870438 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.874522 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.883436 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901155 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d6fb8c-07d0-41f4-b027-76d126425a78-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901203 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901233 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d6fb8c-07d0-41f4-b027-76d126425a78-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901307 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-config\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7d6fb8c-07d0-41f4-b027-76d126425a78-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901350 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zndcn\" (UniqueName: \"kubernetes.io/projected/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-kube-api-access-zndcn\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901389 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901414 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901431 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9clgw\" (UniqueName: \"kubernetes.io/projected/a7d6fb8c-07d0-41f4-b027-76d126425a78-kube-api-access-9clgw\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.901445 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d6fb8c-07d0-41f4-b027-76d126425a78-config\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.902390 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.902440 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d6fb8c-07d0-41f4-b027-76d126425a78-config\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.903213 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7d6fb8c-07d0-41f4-b027-76d126425a78-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.903286 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-config\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.903654 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.903667 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7d6fb8c-07d0-41f4-b027-76d126425a78-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.904280 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.904315 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9098d1964366b0097d75357d1a59610c6e4eae095ada3e84e68c1cd4e1497b9f/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.914000 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.914133 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/620c4a04e8e400a535b164778be4ad98dcbeb842aea6c497ba16b26531cc8f75/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.914567 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d6fb8c-07d0-41f4-b027-76d126425a78-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.914895 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.919656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9clgw\" (UniqueName: \"kubernetes.io/projected/a7d6fb8c-07d0-41f4-b027-76d126425a78-kube-api-access-9clgw\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.923082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zndcn\" (UniqueName: \"kubernetes.io/projected/3ba68c62-f0d6-4472-a7db-9c24b70fbb58-kube-api-access-zndcn\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.952039 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.956082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab5b565c-3ddd-4428-9711-c3a9af899b5e\") pod \"ovsdbserver-sb-2\" (UID: \"3ba68c62-f0d6-4472-a7db-9c24b70fbb58\") " pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.956366 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb7a55de-9d22-49e6-9260-b059d4dd3865\") pod \"ovsdbserver-sb-1\" (UID: \"a7d6fb8c-07d0-41f4-b027-76d126425a78\") " pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.971476 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:46 crc kubenswrapper[4758]: I1203 18:26:46.980077 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.002972 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49778bc2-4f74-4f86-9558-f15d22013db1-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.003023 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51e5a5cf-4500-43de-90bf-788460adcfe5-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.003056 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k75x8\" (UniqueName: \"kubernetes.io/projected/de936fcf-6748-44f0-ad21-e62bec06dba8-kube-api-access-k75x8\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.003084 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.003290 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49778bc2-4f74-4f86-9558-f15d22013db1-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.004497 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de936fcf-6748-44f0-ad21-e62bec06dba8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.004573 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwbp4\" (UniqueName: \"kubernetes.io/projected/51e5a5cf-4500-43de-90bf-788460adcfe5-kube-api-access-rwbp4\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.004638 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de936fcf-6748-44f0-ad21-e62bec06dba8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.004716 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51e5a5cf-4500-43de-90bf-788460adcfe5-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.004737 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjqxw\" (UniqueName: \"kubernetes.io/projected/49778bc2-4f74-4f86-9558-f15d22013db1-kube-api-access-tjqxw\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.004829 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.005611 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e5a5cf-4500-43de-90bf-788460adcfe5-config\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.005640 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e5a5cf-4500-43de-90bf-788460adcfe5-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.005661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49778bc2-4f74-4f86-9558-f15d22013db1-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.005939 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49778bc2-4f74-4f86-9558-f15d22013db1-config\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.006095 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.006180 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de936fcf-6748-44f0-ad21-e62bec06dba8-config\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.006238 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/de936fcf-6748-44f0-ad21-e62bec06dba8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108258 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de936fcf-6748-44f0-ad21-e62bec06dba8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwbp4\" (UniqueName: \"kubernetes.io/projected/51e5a5cf-4500-43de-90bf-788460adcfe5-kube-api-access-rwbp4\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de936fcf-6748-44f0-ad21-e62bec06dba8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51e5a5cf-4500-43de-90bf-788460adcfe5-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108417 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjqxw\" (UniqueName: \"kubernetes.io/projected/49778bc2-4f74-4f86-9558-f15d22013db1-kube-api-access-tjqxw\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108446 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108483 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e5a5cf-4500-43de-90bf-788460adcfe5-config\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108508 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e5a5cf-4500-43de-90bf-788460adcfe5-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108535 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49778bc2-4f74-4f86-9558-f15d22013db1-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108801 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49778bc2-4f74-4f86-9558-f15d22013db1-config\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108830 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108853 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de936fcf-6748-44f0-ad21-e62bec06dba8-config\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108873 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/de936fcf-6748-44f0-ad21-e62bec06dba8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108897 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49778bc2-4f74-4f86-9558-f15d22013db1-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108915 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51e5a5cf-4500-43de-90bf-788460adcfe5-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108945 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k75x8\" (UniqueName: \"kubernetes.io/projected/de936fcf-6748-44f0-ad21-e62bec06dba8-kube-api-access-k75x8\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108971 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.108990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49778bc2-4f74-4f86-9558-f15d22013db1-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.109118 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49778bc2-4f74-4f86-9558-f15d22013db1-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.109460 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/de936fcf-6748-44f0-ad21-e62bec06dba8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.109497 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51e5a5cf-4500-43de-90bf-788460adcfe5-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.109700 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de936fcf-6748-44f0-ad21-e62bec06dba8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.110120 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de936fcf-6748-44f0-ad21-e62bec06dba8-config\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.111200 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49778bc2-4f74-4f86-9558-f15d22013db1-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.111354 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49778bc2-4f74-4f86-9558-f15d22013db1-config\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.112183 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.112222 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f59c6f97373d4f993882fc7d22381cace9df97c044168813aecef85f5b7baa22/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.112604 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.112628 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/52394e6178a769e3b3e6e6fd06c56abac2b3ed08b3c910178170d2cda384cf5f/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.114531 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e5a5cf-4500-43de-90bf-788460adcfe5-config\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.116005 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51e5a5cf-4500-43de-90bf-788460adcfe5-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.116477 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de936fcf-6748-44f0-ad21-e62bec06dba8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.116972 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.117004 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bc8f0a3e1c6ae8e01243c67b289ac4a29c6a9b0443193a5cf5b428edbf0f92d5/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.118259 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49778bc2-4f74-4f86-9558-f15d22013db1-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.118725 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e5a5cf-4500-43de-90bf-788460adcfe5-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.124514 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k75x8\" (UniqueName: \"kubernetes.io/projected/de936fcf-6748-44f0-ad21-e62bec06dba8-kube-api-access-k75x8\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.127791 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjqxw\" (UniqueName: \"kubernetes.io/projected/49778bc2-4f74-4f86-9558-f15d22013db1-kube-api-access-tjqxw\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.131046 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwbp4\" (UniqueName: \"kubernetes.io/projected/51e5a5cf-4500-43de-90bf-788460adcfe5-kube-api-access-rwbp4\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.149625 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7134e1b6-4481-42d9-93c1-1a448c449c0f\") pod \"ovsdbserver-nb-1\" (UID: \"49778bc2-4f74-4f86-9558-f15d22013db1\") " pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.151243 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8fccb5ea-ceef-454c-a495-c253e58ab953\") pod \"ovsdbserver-nb-2\" (UID: \"51e5a5cf-4500-43de-90bf-788460adcfe5\") " pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.153951 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ecd5fcb2-01bf-440f-b93a-1025cff211b9\") pod \"ovsdbserver-nb-0\" (UID: \"de936fcf-6748-44f0-ad21-e62bec06dba8\") " pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.331298 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.346372 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.361055 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.554328 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.634755 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 03 18:26:47 crc kubenswrapper[4758]: I1203 18:26:47.903726 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.005353 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 03 18:26:48 crc kubenswrapper[4758]: W1203 18:26:48.016460 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51e5a5cf_4500_43de_90bf_788460adcfe5.slice/crio-9948d3383ef6102b39729b48026e9a692d36aeabfb100d5f8bcf84921301d1f6 WatchSource:0}: Error finding container 9948d3383ef6102b39729b48026e9a692d36aeabfb100d5f8bcf84921301d1f6: Status 404 returned error can't find the container with id 9948d3383ef6102b39729b48026e9a692d36aeabfb100d5f8bcf84921301d1f6 Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.097745 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"de936fcf-6748-44f0-ad21-e62bec06dba8","Type":"ContainerStarted","Data":"f50b73e9d2967b0b7468f005a4eb66b02c4e971ea4550edcc54dbeb872ded2d3"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.099216 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"51e5a5cf-4500-43de-90bf-788460adcfe5","Type":"ContainerStarted","Data":"9948d3383ef6102b39729b48026e9a692d36aeabfb100d5f8bcf84921301d1f6"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.102006 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"a7d6fb8c-07d0-41f4-b027-76d126425a78","Type":"ContainerStarted","Data":"7a20f6f6a9ce579fecbcf6d56e7e9b6e9caceea51208c5bbca8bd87a4878784d"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.102052 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"a7d6fb8c-07d0-41f4-b027-76d126425a78","Type":"ContainerStarted","Data":"3d2d36e5bef87d51df2cd902f51ea2e2e9d70c2d293a5f43579a745af04d57d5"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.102067 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"a7d6fb8c-07d0-41f4-b027-76d126425a78","Type":"ContainerStarted","Data":"a0a855d82d3482597db02112f4ba57e3c9cef42dae519d081f1517db5056ff95"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.106863 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6812355e-c51f-499b-bfe0-3d78d40e86c7","Type":"ContainerStarted","Data":"c8f08bc37671733f241aa081603baf7e3752125ddd5a411bfa2e35489e4b5f38"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.106895 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6812355e-c51f-499b-bfe0-3d78d40e86c7","Type":"ContainerStarted","Data":"8cf372d2d2c67217fe778df958334e08307a9a6066eb2e1a78ceddd745178aee"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.106909 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6812355e-c51f-499b-bfe0-3d78d40e86c7","Type":"ContainerStarted","Data":"8ace10cbfa8493c246dcebcea2f09552f49bd896e2dc61515715a01b284a6a82"} Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.122060 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.1220465 podStartE2EDuration="3.1220465s" podCreationTimestamp="2025-12-03 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:48.121131344 +0000 UTC m=+5463.322508205" watchObservedRunningTime="2025-12-03 18:26:48.1220465 +0000 UTC m=+5463.323423361" Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.142710 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.142673791 podStartE2EDuration="3.142673791s" podCreationTimestamp="2025-12-03 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:48.139344933 +0000 UTC m=+5463.340721794" watchObservedRunningTime="2025-12-03 18:26:48.142673791 +0000 UTC m=+5463.344050652" Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.300513 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 03 18:26:48 crc kubenswrapper[4758]: W1203 18:26:48.300575 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ba68c62_f0d6_4472_a7db_9c24b70fbb58.slice/crio-995492d25c37e24a8be08e423aed64908a56f5683de482a53567ab76f4686a63 WatchSource:0}: Error finding container 995492d25c37e24a8be08e423aed64908a56f5683de482a53567ab76f4686a63: Status 404 returned error can't find the container with id 995492d25c37e24a8be08e423aed64908a56f5683de482a53567ab76f4686a63 Dec 03 18:26:48 crc kubenswrapper[4758]: I1203 18:26:48.860487 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137428 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"de936fcf-6748-44f0-ad21-e62bec06dba8","Type":"ContainerStarted","Data":"3605e18781ee815d25156b80c3c81374cdf674db6d537b1fc07f9573e5bcd00b"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137826 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"de936fcf-6748-44f0-ad21-e62bec06dba8","Type":"ContainerStarted","Data":"27504b9dc1384714bb90d360de4b107c5e87ba3b8a9f040dbfd94ce686df05ce"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137843 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"3ba68c62-f0d6-4472-a7db-9c24b70fbb58","Type":"ContainerStarted","Data":"d040f2ed2f4380c9bd8ab5bfb9f16ad2c212e2fa81c373b4f943b1eaeed39436"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137858 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"3ba68c62-f0d6-4472-a7db-9c24b70fbb58","Type":"ContainerStarted","Data":"a571f952c3e6e5f06a944ca6f711712be2d946069366a0883f0887005264461e"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137873 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"3ba68c62-f0d6-4472-a7db-9c24b70fbb58","Type":"ContainerStarted","Data":"995492d25c37e24a8be08e423aed64908a56f5683de482a53567ab76f4686a63"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137887 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"51e5a5cf-4500-43de-90bf-788460adcfe5","Type":"ContainerStarted","Data":"f1f81995ac3308a6e3338629016b593f767dc99abca42b1c2dcf4e068ee71471"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137901 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"51e5a5cf-4500-43de-90bf-788460adcfe5","Type":"ContainerStarted","Data":"9a97f2d422fa2bf3ca44c20dd317fd539ac80a785baf364998bbb507239ba196"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137921 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"49778bc2-4f74-4f86-9558-f15d22013db1","Type":"ContainerStarted","Data":"77a91bd16bb0cdd5843755fa5e870e494c05b10a5d14e45e6fda589d52f7f9a2"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.137936 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"49778bc2-4f74-4f86-9558-f15d22013db1","Type":"ContainerStarted","Data":"85a7bb1bbe3b119ecea7d23916ecb4a1f0815abcc3db0854dbc470b656c15ed1"} Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.157332 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.157315701 podStartE2EDuration="4.157315701s" podCreationTimestamp="2025-12-03 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:49.153826687 +0000 UTC m=+5464.355203558" watchObservedRunningTime="2025-12-03 18:26:49.157315701 +0000 UTC m=+5464.358692562" Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.184980 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.184963669 podStartE2EDuration="4.184963669s" podCreationTimestamp="2025-12-03 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:49.174568292 +0000 UTC m=+5464.375945173" watchObservedRunningTime="2025-12-03 18:26:49.184963669 +0000 UTC m=+5464.386340530" Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.195830 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.195813491 podStartE2EDuration="4.195813491s" podCreationTimestamp="2025-12-03 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:49.193379839 +0000 UTC m=+5464.394756690" watchObservedRunningTime="2025-12-03 18:26:49.195813491 +0000 UTC m=+5464.397190352" Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.953385 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.972528 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:49 crc kubenswrapper[4758]: I1203 18:26:49.980838 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:50 crc kubenswrapper[4758]: I1203 18:26:50.150762 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"49778bc2-4f74-4f86-9558-f15d22013db1","Type":"ContainerStarted","Data":"c4bd913466f63ad91ae43720fc7113910f0e3b7633e686d931710ba23daa6ca8"} Dec 03 18:26:50 crc kubenswrapper[4758]: I1203 18:26:50.189767 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=5.189741966 podStartE2EDuration="5.189741966s" podCreationTimestamp="2025-12-03 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:50.180294975 +0000 UTC m=+5465.381671876" watchObservedRunningTime="2025-12-03 18:26:50.189741966 +0000 UTC m=+5465.391118867" Dec 03 18:26:50 crc kubenswrapper[4758]: I1203 18:26:50.332256 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:50 crc kubenswrapper[4758]: I1203 18:26:50.347315 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:50 crc kubenswrapper[4758]: I1203 18:26:50.361477 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:51 crc kubenswrapper[4758]: I1203 18:26:51.952544 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:51 crc kubenswrapper[4758]: I1203 18:26:51.972197 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:51 crc kubenswrapper[4758]: I1203 18:26:51.980265 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:52 crc kubenswrapper[4758]: I1203 18:26:52.332264 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:52 crc kubenswrapper[4758]: I1203 18:26:52.346876 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:52 crc kubenswrapper[4758]: I1203 18:26:52.361599 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.017140 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.032208 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.073819 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.086547 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.126665 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.262740 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.301627 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54dd464559-t667r"] Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.303293 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.306334 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.331394 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54dd464559-t667r"] Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.373092 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.384760 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.420455 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.431068 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.433810 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.438626 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-config\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.438687 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx6xd\" (UniqueName: \"kubernetes.io/projected/83112555-3f25-4819-8a24-891de2a42386-kube-api-access-cx6xd\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.439158 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-ovsdbserver-sb\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.439549 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-dns-svc\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.541652 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-config\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.541718 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx6xd\" (UniqueName: \"kubernetes.io/projected/83112555-3f25-4819-8a24-891de2a42386-kube-api-access-cx6xd\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.541757 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-ovsdbserver-sb\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.541778 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-dns-svc\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.542628 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-dns-svc\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.542626 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-config\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.542726 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-ovsdbserver-sb\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.560661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx6xd\" (UniqueName: \"kubernetes.io/projected/83112555-3f25-4819-8a24-891de2a42386-kube-api-access-cx6xd\") pod \"dnsmasq-dns-54dd464559-t667r\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.633494 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.818008 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dd464559-t667r"] Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.845280 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86c4f5d6c9-xzm5z"] Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.847325 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.850016 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.863852 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86c4f5d6c9-xzm5z"] Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.946880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-dns-svc\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.946918 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-nb\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.946961 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brtfz\" (UniqueName: \"kubernetes.io/projected/0d56e470-87c2-482f-858b-540fb535086e-kube-api-access-brtfz\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.946983 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-config\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:53 crc kubenswrapper[4758]: I1203 18:26:53.947243 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-sb\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.049121 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-sb\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.049223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-dns-svc\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.049246 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-nb\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.049351 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brtfz\" (UniqueName: \"kubernetes.io/projected/0d56e470-87c2-482f-858b-540fb535086e-kube-api-access-brtfz\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.049386 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-config\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.050319 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-nb\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.050352 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-sb\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.050519 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-dns-svc\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.050784 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-config\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.070215 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brtfz\" (UniqueName: \"kubernetes.io/projected/0d56e470-87c2-482f-858b-540fb535086e-kube-api-access-brtfz\") pod \"dnsmasq-dns-86c4f5d6c9-xzm5z\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.163320 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dd464559-t667r"] Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.181921 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.189792 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd464559-t667r" event={"ID":"83112555-3f25-4819-8a24-891de2a42386","Type":"ContainerStarted","Data":"8e1a06435a2fe9c097e4ad744116ac4fcfc21337a54c51ce21540e70a3351bef"} Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.267585 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 03 18:26:54 crc kubenswrapper[4758]: I1203 18:26:54.726917 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86c4f5d6c9-xzm5z"] Dec 03 18:26:54 crc kubenswrapper[4758]: W1203 18:26:54.732609 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d56e470_87c2_482f_858b_540fb535086e.slice/crio-5b85f5131f9aa55d102bbff77d488dc802ac8088382f5f9c15e97026bc6cd705 WatchSource:0}: Error finding container 5b85f5131f9aa55d102bbff77d488dc802ac8088382f5f9c15e97026bc6cd705: Status 404 returned error can't find the container with id 5b85f5131f9aa55d102bbff77d488dc802ac8088382f5f9c15e97026bc6cd705 Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.200268 4758 generic.go:334] "Generic (PLEG): container finished" podID="0d56e470-87c2-482f-858b-540fb535086e" containerID="a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb" exitCode=0 Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.200369 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" event={"ID":"0d56e470-87c2-482f-858b-540fb535086e","Type":"ContainerDied","Data":"a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb"} Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.200404 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" event={"ID":"0d56e470-87c2-482f-858b-540fb535086e","Type":"ContainerStarted","Data":"5b85f5131f9aa55d102bbff77d488dc802ac8088382f5f9c15e97026bc6cd705"} Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.203385 4758 generic.go:334] "Generic (PLEG): container finished" podID="83112555-3f25-4819-8a24-891de2a42386" containerID="1b3e726b5cc70760993651e1129f251d23083df2fd66f80afc6a7c08cd87b505" exitCode=0 Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.203559 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd464559-t667r" event={"ID":"83112555-3f25-4819-8a24-891de2a42386","Type":"ContainerDied","Data":"1b3e726b5cc70760993651e1129f251d23083df2fd66f80afc6a7c08cd87b505"} Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.510170 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.591371 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-ovsdbserver-sb\") pod \"83112555-3f25-4819-8a24-891de2a42386\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.591411 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx6xd\" (UniqueName: \"kubernetes.io/projected/83112555-3f25-4819-8a24-891de2a42386-kube-api-access-cx6xd\") pod \"83112555-3f25-4819-8a24-891de2a42386\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.591429 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-config\") pod \"83112555-3f25-4819-8a24-891de2a42386\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.591462 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-dns-svc\") pod \"83112555-3f25-4819-8a24-891de2a42386\" (UID: \"83112555-3f25-4819-8a24-891de2a42386\") " Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.595825 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83112555-3f25-4819-8a24-891de2a42386-kube-api-access-cx6xd" (OuterVolumeSpecName: "kube-api-access-cx6xd") pod "83112555-3f25-4819-8a24-891de2a42386" (UID: "83112555-3f25-4819-8a24-891de2a42386"). InnerVolumeSpecName "kube-api-access-cx6xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.611341 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83112555-3f25-4819-8a24-891de2a42386" (UID: "83112555-3f25-4819-8a24-891de2a42386"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.625707 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-config" (OuterVolumeSpecName: "config") pod "83112555-3f25-4819-8a24-891de2a42386" (UID: "83112555-3f25-4819-8a24-891de2a42386"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.629830 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83112555-3f25-4819-8a24-891de2a42386" (UID: "83112555-3f25-4819-8a24-891de2a42386"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.694145 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.694190 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.694214 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx6xd\" (UniqueName: \"kubernetes.io/projected/83112555-3f25-4819-8a24-891de2a42386-kube-api-access-cx6xd\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:55 crc kubenswrapper[4758]: I1203 18:26:55.694232 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83112555-3f25-4819-8a24-891de2a42386-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.226540 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd464559-t667r" event={"ID":"83112555-3f25-4819-8a24-891de2a42386","Type":"ContainerDied","Data":"8e1a06435a2fe9c097e4ad744116ac4fcfc21337a54c51ce21540e70a3351bef"} Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.226997 4758 scope.go:117] "RemoveContainer" containerID="1b3e726b5cc70760993651e1129f251d23083df2fd66f80afc6a7c08cd87b505" Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.226552 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd464559-t667r" Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.233111 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" event={"ID":"0d56e470-87c2-482f-858b-540fb535086e","Type":"ContainerStarted","Data":"ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be"} Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.233302 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.259476 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" podStartSLOduration=3.259457827 podStartE2EDuration="3.259457827s" podCreationTimestamp="2025-12-03 18:26:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:56.255438478 +0000 UTC m=+5471.456815349" watchObservedRunningTime="2025-12-03 18:26:56.259457827 +0000 UTC m=+5471.460834688" Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.323557 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dd464559-t667r"] Dec 03 18:26:56 crc kubenswrapper[4758]: I1203 18:26:56.329326 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54dd464559-t667r"] Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.127297 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83112555-3f25-4819-8a24-891de2a42386" path="/var/lib/kubelet/pods/83112555-3f25-4819-8a24-891de2a42386/volumes" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.513310 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 03 18:26:57 crc kubenswrapper[4758]: E1203 18:26:57.513702 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83112555-3f25-4819-8a24-891de2a42386" containerName="init" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.513715 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="83112555-3f25-4819-8a24-891de2a42386" containerName="init" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.513871 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="83112555-3f25-4819-8a24-891de2a42386" containerName="init" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.514382 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.523617 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.529769 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.624591 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5baafff9-4b8b-4714-8758-87fa113058a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.624756 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnxb2\" (UniqueName: \"kubernetes.io/projected/3206afc1-0965-4c69-a90c-95be3a0605b3-kube-api-access-pnxb2\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.624862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3206afc1-0965-4c69-a90c-95be3a0605b3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.726093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3206afc1-0965-4c69-a90c-95be3a0605b3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.726549 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5baafff9-4b8b-4714-8758-87fa113058a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.726665 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnxb2\" (UniqueName: \"kubernetes.io/projected/3206afc1-0965-4c69-a90c-95be3a0605b3-kube-api-access-pnxb2\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.731095 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.731141 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5baafff9-4b8b-4714-8758-87fa113058a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1d396016f64558c1faf4dddd6680f8a0c5040bdfc18f5e30fbff04423eccf0d9/globalmount\"" pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.737371 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3206afc1-0965-4c69-a90c-95be3a0605b3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.762033 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnxb2\" (UniqueName: \"kubernetes.io/projected/3206afc1-0965-4c69-a90c-95be3a0605b3-kube-api-access-pnxb2\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.789624 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5baafff9-4b8b-4714-8758-87fa113058a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") pod \"ovn-copy-data\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " pod="openstack/ovn-copy-data" Dec 03 18:26:57 crc kubenswrapper[4758]: I1203 18:26:57.848714 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 03 18:26:58 crc kubenswrapper[4758]: I1203 18:26:58.377877 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 03 18:26:59 crc kubenswrapper[4758]: I1203 18:26:59.278985 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3206afc1-0965-4c69-a90c-95be3a0605b3","Type":"ContainerStarted","Data":"37a9efb4e5ef7c6a7fb36e6cf8cba641b6971768e8011861d63f186d28ee2af3"} Dec 03 18:26:59 crc kubenswrapper[4758]: I1203 18:26:59.279347 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3206afc1-0965-4c69-a90c-95be3a0605b3","Type":"ContainerStarted","Data":"13912022e8ffef461395dfc9683b99e8d48dc59178c2a16f929d00ce3cb93a5a"} Dec 03 18:26:59 crc kubenswrapper[4758]: I1203 18:26:59.304674 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.30465027 podStartE2EDuration="3.30465027s" podCreationTimestamp="2025-12-03 18:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:26:59.301225269 +0000 UTC m=+5474.502602170" watchObservedRunningTime="2025-12-03 18:26:59.30465027 +0000 UTC m=+5474.506027171" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.182846 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.282301 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-tsjb2"] Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.282642 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerName="dnsmasq-dns" containerID="cri-o://fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6" gracePeriod=10 Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.745961 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.848409 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-dns-svc\") pod \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.848659 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-config\") pod \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.848800 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f62nk\" (UniqueName: \"kubernetes.io/projected/366ef7ee-bf1d-4570-b3c0-da98dccf2685-kube-api-access-f62nk\") pod \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\" (UID: \"366ef7ee-bf1d-4570-b3c0-da98dccf2685\") " Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.856752 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366ef7ee-bf1d-4570-b3c0-da98dccf2685-kube-api-access-f62nk" (OuterVolumeSpecName: "kube-api-access-f62nk") pod "366ef7ee-bf1d-4570-b3c0-da98dccf2685" (UID: "366ef7ee-bf1d-4570-b3c0-da98dccf2685"). InnerVolumeSpecName "kube-api-access-f62nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.890507 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "366ef7ee-bf1d-4570-b3c0-da98dccf2685" (UID: "366ef7ee-bf1d-4570-b3c0-da98dccf2685"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.897200 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-config" (OuterVolumeSpecName: "config") pod "366ef7ee-bf1d-4570-b3c0-da98dccf2685" (UID: "366ef7ee-bf1d-4570-b3c0-da98dccf2685"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.951620 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.951702 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f62nk\" (UniqueName: \"kubernetes.io/projected/366ef7ee-bf1d-4570-b3c0-da98dccf2685-kube-api-access-f62nk\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:04 crc kubenswrapper[4758]: I1203 18:27:04.951715 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ef7ee-bf1d-4570-b3c0-da98dccf2685-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.071428 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 18:27:05 crc kubenswrapper[4758]: E1203 18:27:05.072191 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerName="dnsmasq-dns" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.072218 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerName="dnsmasq-dns" Dec 03 18:27:05 crc kubenswrapper[4758]: E1203 18:27:05.072247 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerName="init" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.072254 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerName="init" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.072475 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerName="dnsmasq-dns" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.079814 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.089824 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.089860 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.090565 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-wvr6b" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.106011 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.160190 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.160241 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-scripts\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.160263 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.160285 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-config\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.160314 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsknq\" (UniqueName: \"kubernetes.io/projected/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-kube-api-access-nsknq\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.261948 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.262001 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-scripts\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.262030 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.262053 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-config\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.262081 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsknq\" (UniqueName: \"kubernetes.io/projected/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-kube-api-access-nsknq\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.263853 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.263917 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-scripts\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.264433 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-config\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.269076 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.281492 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsknq\" (UniqueName: \"kubernetes.io/projected/41e72dc3-c313-42dd-a31e-c7bbcf63ff5a-kube-api-access-nsknq\") pod \"ovn-northd-0\" (UID: \"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a\") " pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.334808 4758 generic.go:334] "Generic (PLEG): container finished" podID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" containerID="fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6" exitCode=0 Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.334868 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" event={"ID":"366ef7ee-bf1d-4570-b3c0-da98dccf2685","Type":"ContainerDied","Data":"fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6"} Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.334906 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" event={"ID":"366ef7ee-bf1d-4570-b3c0-da98dccf2685","Type":"ContainerDied","Data":"053099eff7aa2c0180bee02dd4a43d0898cc9d1a44cbe6d90ae2094a2698f30d"} Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.334922 4758 scope.go:117] "RemoveContainer" containerID="fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.335109 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-tsjb2" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.361834 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-tsjb2"] Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.369450 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-tsjb2"] Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.372059 4758 scope.go:117] "RemoveContainer" containerID="bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.393667 4758 scope.go:117] "RemoveContainer" containerID="fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6" Dec 03 18:27:05 crc kubenswrapper[4758]: E1203 18:27:05.394205 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6\": container with ID starting with fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6 not found: ID does not exist" containerID="fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.394233 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6"} err="failed to get container status \"fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6\": rpc error: code = NotFound desc = could not find container \"fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6\": container with ID starting with fd7ae4ecec9b5d9db9f8ac5306a028716770942ac1238d6dfc5c44315c899df6 not found: ID does not exist" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.394254 4758 scope.go:117] "RemoveContainer" containerID="bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014" Dec 03 18:27:05 crc kubenswrapper[4758]: E1203 18:27:05.394506 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014\": container with ID starting with bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014 not found: ID does not exist" containerID="bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.394545 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014"} err="failed to get container status \"bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014\": rpc error: code = NotFound desc = could not find container \"bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014\": container with ID starting with bccde4837a2406ed8a0e4cd8ba191acc37333f8b83d22b989230bcb1de580014 not found: ID does not exist" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.412419 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 18:27:05 crc kubenswrapper[4758]: I1203 18:27:05.918332 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 18:27:06 crc kubenswrapper[4758]: I1203 18:27:06.348201 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a","Type":"ContainerStarted","Data":"82c2465a87e7f80a3ded700937a7e9b2c5abb9bc085c8d24b4f25351b9a756c1"} Dec 03 18:27:06 crc kubenswrapper[4758]: I1203 18:27:06.348858 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 18:27:06 crc kubenswrapper[4758]: I1203 18:27:06.348896 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a","Type":"ContainerStarted","Data":"cca6ed45924726a75c84b7865229d988785b55b9355bc9d7d3cf8590757606a2"} Dec 03 18:27:06 crc kubenswrapper[4758]: I1203 18:27:06.348916 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"41e72dc3-c313-42dd-a31e-c7bbcf63ff5a","Type":"ContainerStarted","Data":"88f1aebdc08fbd818b4809a89b85b46f35f996e877562cfe6be2a43e6bd5dc2f"} Dec 03 18:27:06 crc kubenswrapper[4758]: I1203 18:27:06.371269 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.371239005 podStartE2EDuration="1.371239005s" podCreationTimestamp="2025-12-03 18:27:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:27:06.362857117 +0000 UTC m=+5481.564234008" watchObservedRunningTime="2025-12-03 18:27:06.371239005 +0000 UTC m=+5481.572615906" Dec 03 18:27:07 crc kubenswrapper[4758]: I1203 18:27:07.128802 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366ef7ee-bf1d-4570-b3c0-da98dccf2685" path="/var/lib/kubelet/pods/366ef7ee-bf1d-4570-b3c0-da98dccf2685/volumes" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.842264 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jvjf4"] Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.843716 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.856476 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jvjf4"] Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.922581 4758 scope.go:117] "RemoveContainer" containerID="71b523946e959456b2486c94a4a693522116bad2293b7b72081619f367467920" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.948050 4758 scope.go:117] "RemoveContainer" containerID="1d2f671bd647a86d14ca5e8bfec6b0b60d9535f422bafd64df030c37f2816ed2" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.955522 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8074-account-create-update-9cjh7"] Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.956987 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.959133 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.962982 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8074-account-create-update-9cjh7"] Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.970197 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53cff074-8895-4fd4-90f3-eebeea7c9caa-operator-scripts\") pod \"keystone-db-create-jvjf4\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:10 crc kubenswrapper[4758]: I1203 18:27:10.970353 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9572s\" (UniqueName: \"kubernetes.io/projected/53cff074-8895-4fd4-90f3-eebeea7c9caa-kube-api-access-9572s\") pod \"keystone-db-create-jvjf4\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.071523 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9572s\" (UniqueName: \"kubernetes.io/projected/53cff074-8895-4fd4-90f3-eebeea7c9caa-kube-api-access-9572s\") pod \"keystone-db-create-jvjf4\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.071589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm5pm\" (UniqueName: \"kubernetes.io/projected/2edbbc54-5da0-4483-bff6-6f8c919e9db2-kube-api-access-sm5pm\") pod \"keystone-8074-account-create-update-9cjh7\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.071631 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53cff074-8895-4fd4-90f3-eebeea7c9caa-operator-scripts\") pod \"keystone-db-create-jvjf4\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.071668 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2edbbc54-5da0-4483-bff6-6f8c919e9db2-operator-scripts\") pod \"keystone-8074-account-create-update-9cjh7\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.072804 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53cff074-8895-4fd4-90f3-eebeea7c9caa-operator-scripts\") pod \"keystone-db-create-jvjf4\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.087909 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9572s\" (UniqueName: \"kubernetes.io/projected/53cff074-8895-4fd4-90f3-eebeea7c9caa-kube-api-access-9572s\") pod \"keystone-db-create-jvjf4\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.169012 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.172653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2edbbc54-5da0-4483-bff6-6f8c919e9db2-operator-scripts\") pod \"keystone-8074-account-create-update-9cjh7\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.172840 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm5pm\" (UniqueName: \"kubernetes.io/projected/2edbbc54-5da0-4483-bff6-6f8c919e9db2-kube-api-access-sm5pm\") pod \"keystone-8074-account-create-update-9cjh7\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.173436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2edbbc54-5da0-4483-bff6-6f8c919e9db2-operator-scripts\") pod \"keystone-8074-account-create-update-9cjh7\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.189658 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm5pm\" (UniqueName: \"kubernetes.io/projected/2edbbc54-5da0-4483-bff6-6f8c919e9db2-kube-api-access-sm5pm\") pod \"keystone-8074-account-create-update-9cjh7\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.336344 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.587460 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jvjf4"] Dec 03 18:27:11 crc kubenswrapper[4758]: W1203 18:27:11.588322 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53cff074_8895_4fd4_90f3_eebeea7c9caa.slice/crio-c5614409b7634befc5288a1f54ca4d36c1fbc0925fe48d24e883cf737626417a WatchSource:0}: Error finding container c5614409b7634befc5288a1f54ca4d36c1fbc0925fe48d24e883cf737626417a: Status 404 returned error can't find the container with id c5614409b7634befc5288a1f54ca4d36c1fbc0925fe48d24e883cf737626417a Dec 03 18:27:11 crc kubenswrapper[4758]: I1203 18:27:11.758113 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8074-account-create-update-9cjh7"] Dec 03 18:27:11 crc kubenswrapper[4758]: W1203 18:27:11.760108 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2edbbc54_5da0_4483_bff6_6f8c919e9db2.slice/crio-7837e3776cd16502c5a6a2cf47b1c8ab6aaca3ccbb91396c378c35e2d5f2a3a2 WatchSource:0}: Error finding container 7837e3776cd16502c5a6a2cf47b1c8ab6aaca3ccbb91396c378c35e2d5f2a3a2: Status 404 returned error can't find the container with id 7837e3776cd16502c5a6a2cf47b1c8ab6aaca3ccbb91396c378c35e2d5f2a3a2 Dec 03 18:27:12 crc kubenswrapper[4758]: I1203 18:27:12.410934 4758 generic.go:334] "Generic (PLEG): container finished" podID="53cff074-8895-4fd4-90f3-eebeea7c9caa" containerID="b7da49a4b675859df73d5bd11925599345a99ffc4e7ac8dadfb365a87b1c418b" exitCode=0 Dec 03 18:27:12 crc kubenswrapper[4758]: I1203 18:27:12.411054 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jvjf4" event={"ID":"53cff074-8895-4fd4-90f3-eebeea7c9caa","Type":"ContainerDied","Data":"b7da49a4b675859df73d5bd11925599345a99ffc4e7ac8dadfb365a87b1c418b"} Dec 03 18:27:12 crc kubenswrapper[4758]: I1203 18:27:12.411438 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jvjf4" event={"ID":"53cff074-8895-4fd4-90f3-eebeea7c9caa","Type":"ContainerStarted","Data":"c5614409b7634befc5288a1f54ca4d36c1fbc0925fe48d24e883cf737626417a"} Dec 03 18:27:12 crc kubenswrapper[4758]: I1203 18:27:12.414005 4758 generic.go:334] "Generic (PLEG): container finished" podID="2edbbc54-5da0-4483-bff6-6f8c919e9db2" containerID="dc885c069b333d8660a1941fc0f26fe6797ee8d95297a047e5dac1ab6649b277" exitCode=0 Dec 03 18:27:12 crc kubenswrapper[4758]: I1203 18:27:12.414063 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8074-account-create-update-9cjh7" event={"ID":"2edbbc54-5da0-4483-bff6-6f8c919e9db2","Type":"ContainerDied","Data":"dc885c069b333d8660a1941fc0f26fe6797ee8d95297a047e5dac1ab6649b277"} Dec 03 18:27:12 crc kubenswrapper[4758]: I1203 18:27:12.414104 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8074-account-create-update-9cjh7" event={"ID":"2edbbc54-5da0-4483-bff6-6f8c919e9db2","Type":"ContainerStarted","Data":"7837e3776cd16502c5a6a2cf47b1c8ab6aaca3ccbb91396c378c35e2d5f2a3a2"} Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.872075 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.877862 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.967568 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm5pm\" (UniqueName: \"kubernetes.io/projected/2edbbc54-5da0-4483-bff6-6f8c919e9db2-kube-api-access-sm5pm\") pod \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.967697 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53cff074-8895-4fd4-90f3-eebeea7c9caa-operator-scripts\") pod \"53cff074-8895-4fd4-90f3-eebeea7c9caa\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.967785 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9572s\" (UniqueName: \"kubernetes.io/projected/53cff074-8895-4fd4-90f3-eebeea7c9caa-kube-api-access-9572s\") pod \"53cff074-8895-4fd4-90f3-eebeea7c9caa\" (UID: \"53cff074-8895-4fd4-90f3-eebeea7c9caa\") " Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.967857 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2edbbc54-5da0-4483-bff6-6f8c919e9db2-operator-scripts\") pod \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\" (UID: \"2edbbc54-5da0-4483-bff6-6f8c919e9db2\") " Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.969841 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53cff074-8895-4fd4-90f3-eebeea7c9caa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53cff074-8895-4fd4-90f3-eebeea7c9caa" (UID: "53cff074-8895-4fd4-90f3-eebeea7c9caa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.974806 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2edbbc54-5da0-4483-bff6-6f8c919e9db2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2edbbc54-5da0-4483-bff6-6f8c919e9db2" (UID: "2edbbc54-5da0-4483-bff6-6f8c919e9db2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.975436 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53cff074-8895-4fd4-90f3-eebeea7c9caa-kube-api-access-9572s" (OuterVolumeSpecName: "kube-api-access-9572s") pod "53cff074-8895-4fd4-90f3-eebeea7c9caa" (UID: "53cff074-8895-4fd4-90f3-eebeea7c9caa"). InnerVolumeSpecName "kube-api-access-9572s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:13 crc kubenswrapper[4758]: I1203 18:27:13.985276 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2edbbc54-5da0-4483-bff6-6f8c919e9db2-kube-api-access-sm5pm" (OuterVolumeSpecName: "kube-api-access-sm5pm") pod "2edbbc54-5da0-4483-bff6-6f8c919e9db2" (UID: "2edbbc54-5da0-4483-bff6-6f8c919e9db2"). InnerVolumeSpecName "kube-api-access-sm5pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.069635 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm5pm\" (UniqueName: \"kubernetes.io/projected/2edbbc54-5da0-4483-bff6-6f8c919e9db2-kube-api-access-sm5pm\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.069669 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53cff074-8895-4fd4-90f3-eebeea7c9caa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.069718 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9572s\" (UniqueName: \"kubernetes.io/projected/53cff074-8895-4fd4-90f3-eebeea7c9caa-kube-api-access-9572s\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.069728 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2edbbc54-5da0-4483-bff6-6f8c919e9db2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.431737 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8074-account-create-update-9cjh7" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.431731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8074-account-create-update-9cjh7" event={"ID":"2edbbc54-5da0-4483-bff6-6f8c919e9db2","Type":"ContainerDied","Data":"7837e3776cd16502c5a6a2cf47b1c8ab6aaca3ccbb91396c378c35e2d5f2a3a2"} Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.431890 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7837e3776cd16502c5a6a2cf47b1c8ab6aaca3ccbb91396c378c35e2d5f2a3a2" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.434659 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jvjf4" event={"ID":"53cff074-8895-4fd4-90f3-eebeea7c9caa","Type":"ContainerDied","Data":"c5614409b7634befc5288a1f54ca4d36c1fbc0925fe48d24e883cf737626417a"} Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.434723 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5614409b7634befc5288a1f54ca4d36c1fbc0925fe48d24e883cf737626417a" Dec 03 18:27:14 crc kubenswrapper[4758]: I1203 18:27:14.434761 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jvjf4" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.400506 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-rpgg5"] Dec 03 18:27:16 crc kubenswrapper[4758]: E1203 18:27:16.401160 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cff074-8895-4fd4-90f3-eebeea7c9caa" containerName="mariadb-database-create" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.401176 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cff074-8895-4fd4-90f3-eebeea7c9caa" containerName="mariadb-database-create" Dec 03 18:27:16 crc kubenswrapper[4758]: E1203 18:27:16.401213 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edbbc54-5da0-4483-bff6-6f8c919e9db2" containerName="mariadb-account-create-update" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.401221 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edbbc54-5da0-4483-bff6-6f8c919e9db2" containerName="mariadb-account-create-update" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.401407 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edbbc54-5da0-4483-bff6-6f8c919e9db2" containerName="mariadb-account-create-update" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.401434 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="53cff074-8895-4fd4-90f3-eebeea7c9caa" containerName="mariadb-database-create" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.402086 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.404293 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.404922 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.406924 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gp4jx" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.408229 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-combined-ca-bundle\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.408782 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-config-data\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.408956 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzjzv\" (UniqueName: \"kubernetes.io/projected/186e38e4-ec80-424f-bf8e-745158c88b21-kube-api-access-hzjzv\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.410257 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.416624 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rpgg5"] Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.509829 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-config-data\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.509893 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzjzv\" (UniqueName: \"kubernetes.io/projected/186e38e4-ec80-424f-bf8e-745158c88b21-kube-api-access-hzjzv\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.509946 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-combined-ca-bundle\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.517273 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-combined-ca-bundle\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.517413 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-config-data\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.526094 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzjzv\" (UniqueName: \"kubernetes.io/projected/186e38e4-ec80-424f-bf8e-745158c88b21-kube-api-access-hzjzv\") pod \"keystone-db-sync-rpgg5\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:16 crc kubenswrapper[4758]: I1203 18:27:16.775926 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:17 crc kubenswrapper[4758]: I1203 18:27:17.205071 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rpgg5"] Dec 03 18:27:17 crc kubenswrapper[4758]: W1203 18:27:17.207837 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod186e38e4_ec80_424f_bf8e_745158c88b21.slice/crio-37b641f44d6c90516c0a13de812c49f75e5e4f7498a7ffc3bfef7ee980c1634f WatchSource:0}: Error finding container 37b641f44d6c90516c0a13de812c49f75e5e4f7498a7ffc3bfef7ee980c1634f: Status 404 returned error can't find the container with id 37b641f44d6c90516c0a13de812c49f75e5e4f7498a7ffc3bfef7ee980c1634f Dec 03 18:27:17 crc kubenswrapper[4758]: I1203 18:27:17.481988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rpgg5" event={"ID":"186e38e4-ec80-424f-bf8e-745158c88b21","Type":"ContainerStarted","Data":"37b641f44d6c90516c0a13de812c49f75e5e4f7498a7ffc3bfef7ee980c1634f"} Dec 03 18:27:18 crc kubenswrapper[4758]: I1203 18:27:18.496499 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rpgg5" event={"ID":"186e38e4-ec80-424f-bf8e-745158c88b21","Type":"ContainerStarted","Data":"dd15f4692cb8f037a9109df8374e6951f5eea9cdf439455644da5286f779d43c"} Dec 03 18:27:18 crc kubenswrapper[4758]: I1203 18:27:18.535428 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-rpgg5" podStartSLOduration=2.5354032330000003 podStartE2EDuration="2.535403233s" podCreationTimestamp="2025-12-03 18:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:27:18.523016856 +0000 UTC m=+5493.724393717" watchObservedRunningTime="2025-12-03 18:27:18.535403233 +0000 UTC m=+5493.736780104" Dec 03 18:27:19 crc kubenswrapper[4758]: I1203 18:27:19.506338 4758 generic.go:334] "Generic (PLEG): container finished" podID="186e38e4-ec80-424f-bf8e-745158c88b21" containerID="dd15f4692cb8f037a9109df8374e6951f5eea9cdf439455644da5286f779d43c" exitCode=0 Dec 03 18:27:19 crc kubenswrapper[4758]: I1203 18:27:19.506445 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rpgg5" event={"ID":"186e38e4-ec80-424f-bf8e-745158c88b21","Type":"ContainerDied","Data":"dd15f4692cb8f037a9109df8374e6951f5eea9cdf439455644da5286f779d43c"} Dec 03 18:27:20 crc kubenswrapper[4758]: I1203 18:27:20.468118 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 18:27:20 crc kubenswrapper[4758]: I1203 18:27:20.852616 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.003012 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-config-data\") pod \"186e38e4-ec80-424f-bf8e-745158c88b21\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.003253 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-combined-ca-bundle\") pod \"186e38e4-ec80-424f-bf8e-745158c88b21\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.003305 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzjzv\" (UniqueName: \"kubernetes.io/projected/186e38e4-ec80-424f-bf8e-745158c88b21-kube-api-access-hzjzv\") pod \"186e38e4-ec80-424f-bf8e-745158c88b21\" (UID: \"186e38e4-ec80-424f-bf8e-745158c88b21\") " Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.013522 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/186e38e4-ec80-424f-bf8e-745158c88b21-kube-api-access-hzjzv" (OuterVolumeSpecName: "kube-api-access-hzjzv") pod "186e38e4-ec80-424f-bf8e-745158c88b21" (UID: "186e38e4-ec80-424f-bf8e-745158c88b21"). InnerVolumeSpecName "kube-api-access-hzjzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.026534 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "186e38e4-ec80-424f-bf8e-745158c88b21" (UID: "186e38e4-ec80-424f-bf8e-745158c88b21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.050934 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-config-data" (OuterVolumeSpecName: "config-data") pod "186e38e4-ec80-424f-bf8e-745158c88b21" (UID: "186e38e4-ec80-424f-bf8e-745158c88b21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.105627 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.105669 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186e38e4-ec80-424f-bf8e-745158c88b21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.105709 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzjzv\" (UniqueName: \"kubernetes.io/projected/186e38e4-ec80-424f-bf8e-745158c88b21-kube-api-access-hzjzv\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.522639 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rpgg5" event={"ID":"186e38e4-ec80-424f-bf8e-745158c88b21","Type":"ContainerDied","Data":"37b641f44d6c90516c0a13de812c49f75e5e4f7498a7ffc3bfef7ee980c1634f"} Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.522694 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rpgg5" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.522709 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b641f44d6c90516c0a13de812c49f75e5e4f7498a7ffc3bfef7ee980c1634f" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.766981 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wmx87"] Dec 03 18:27:21 crc kubenswrapper[4758]: E1203 18:27:21.770209 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="186e38e4-ec80-424f-bf8e-745158c88b21" containerName="keystone-db-sync" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.770242 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="186e38e4-ec80-424f-bf8e-745158c88b21" containerName="keystone-db-sync" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.770511 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="186e38e4-ec80-424f-bf8e-745158c88b21" containerName="keystone-db-sync" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.771270 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.773965 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.774190 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.774384 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gp4jx" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.774493 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.774589 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.777268 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wmx87"] Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.808780 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67674546dc-bp5pk"] Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.812606 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.820540 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67674546dc-bp5pk"] Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.920597 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-fernet-keys\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.920653 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-combined-ca-bundle\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.920763 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-credential-keys\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.920813 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-dns-svc\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.920930 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-config\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.920975 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-sb\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.921027 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-config-data\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.921053 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-scripts\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.921074 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz588\" (UniqueName: \"kubernetes.io/projected/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-kube-api-access-rz588\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.921135 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-nb\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:21 crc kubenswrapper[4758]: I1203 18:27:21.921172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm9vz\" (UniqueName: \"kubernetes.io/projected/1513698d-bd5c-41e8-92e3-cd9ee11ad479-kube-api-access-hm9vz\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022204 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-fernet-keys\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022248 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-combined-ca-bundle\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022306 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-credential-keys\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022348 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-dns-svc\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022388 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-config\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022415 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-sb\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022450 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-config-data\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022474 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-scripts\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022498 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz588\" (UniqueName: \"kubernetes.io/projected/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-kube-api-access-rz588\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022530 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-nb\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.022557 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm9vz\" (UniqueName: \"kubernetes.io/projected/1513698d-bd5c-41e8-92e3-cd9ee11ad479-kube-api-access-hm9vz\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.023632 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-config\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.023720 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-sb\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.023743 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-nb\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.024053 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-dns-svc\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.027388 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-combined-ca-bundle\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.027825 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-scripts\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.028008 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-credential-keys\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.029173 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-fernet-keys\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.033061 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-config-data\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.047673 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz588\" (UniqueName: \"kubernetes.io/projected/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-kube-api-access-rz588\") pod \"dnsmasq-dns-67674546dc-bp5pk\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.048249 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm9vz\" (UniqueName: \"kubernetes.io/projected/1513698d-bd5c-41e8-92e3-cd9ee11ad479-kube-api-access-hm9vz\") pod \"keystone-bootstrap-wmx87\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.097023 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.131806 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.596847 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wmx87"] Dec 03 18:27:22 crc kubenswrapper[4758]: W1203 18:27:22.601976 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1513698d_bd5c_41e8_92e3_cd9ee11ad479.slice/crio-de1a3526623ad21ed1d5dca6e4b5f700c83e111d532b492f7cd2ff31b24720d2 WatchSource:0}: Error finding container de1a3526623ad21ed1d5dca6e4b5f700c83e111d532b492f7cd2ff31b24720d2: Status 404 returned error can't find the container with id de1a3526623ad21ed1d5dca6e4b5f700c83e111d532b492f7cd2ff31b24720d2 Dec 03 18:27:22 crc kubenswrapper[4758]: I1203 18:27:22.642499 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67674546dc-bp5pk"] Dec 03 18:27:22 crc kubenswrapper[4758]: W1203 18:27:22.653999 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9aad7c6_1569_4524_9f1e_c8240fd4ad51.slice/crio-ed33543e41eebfd131e4f973991e9696f149d7c1d3376e6b9b6fd88bfcd1ff61 WatchSource:0}: Error finding container ed33543e41eebfd131e4f973991e9696f149d7c1d3376e6b9b6fd88bfcd1ff61: Status 404 returned error can't find the container with id ed33543e41eebfd131e4f973991e9696f149d7c1d3376e6b9b6fd88bfcd1ff61 Dec 03 18:27:23 crc kubenswrapper[4758]: I1203 18:27:23.583781 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wmx87" event={"ID":"1513698d-bd5c-41e8-92e3-cd9ee11ad479","Type":"ContainerStarted","Data":"9d51de49bcf89ec563eabbe282e0f7116aaeae6a4093531e584b68bdc46b22a9"} Dec 03 18:27:23 crc kubenswrapper[4758]: I1203 18:27:23.583839 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wmx87" event={"ID":"1513698d-bd5c-41e8-92e3-cd9ee11ad479","Type":"ContainerStarted","Data":"de1a3526623ad21ed1d5dca6e4b5f700c83e111d532b492f7cd2ff31b24720d2"} Dec 03 18:27:23 crc kubenswrapper[4758]: I1203 18:27:23.586183 4758 generic.go:334] "Generic (PLEG): container finished" podID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerID="52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417" exitCode=0 Dec 03 18:27:23 crc kubenswrapper[4758]: I1203 18:27:23.586243 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" event={"ID":"f9aad7c6-1569-4524-9f1e-c8240fd4ad51","Type":"ContainerDied","Data":"52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417"} Dec 03 18:27:23 crc kubenswrapper[4758]: I1203 18:27:23.586280 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" event={"ID":"f9aad7c6-1569-4524-9f1e-c8240fd4ad51","Type":"ContainerStarted","Data":"ed33543e41eebfd131e4f973991e9696f149d7c1d3376e6b9b6fd88bfcd1ff61"} Dec 03 18:27:23 crc kubenswrapper[4758]: I1203 18:27:23.601007 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wmx87" podStartSLOduration=2.600987368 podStartE2EDuration="2.600987368s" podCreationTimestamp="2025-12-03 18:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:27:23.598750302 +0000 UTC m=+5498.800127263" watchObservedRunningTime="2025-12-03 18:27:23.600987368 +0000 UTC m=+5498.802364239" Dec 03 18:27:24 crc kubenswrapper[4758]: I1203 18:27:24.596182 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" event={"ID":"f9aad7c6-1569-4524-9f1e-c8240fd4ad51","Type":"ContainerStarted","Data":"48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08"} Dec 03 18:27:24 crc kubenswrapper[4758]: I1203 18:27:24.624278 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" podStartSLOduration=3.624254611 podStartE2EDuration="3.624254611s" podCreationTimestamp="2025-12-03 18:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:27:24.616500192 +0000 UTC m=+5499.817877053" watchObservedRunningTime="2025-12-03 18:27:24.624254611 +0000 UTC m=+5499.825631482" Dec 03 18:27:25 crc kubenswrapper[4758]: I1203 18:27:25.607003 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:26 crc kubenswrapper[4758]: I1203 18:27:26.615320 4758 generic.go:334] "Generic (PLEG): container finished" podID="1513698d-bd5c-41e8-92e3-cd9ee11ad479" containerID="9d51de49bcf89ec563eabbe282e0f7116aaeae6a4093531e584b68bdc46b22a9" exitCode=0 Dec 03 18:27:26 crc kubenswrapper[4758]: I1203 18:27:26.615504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wmx87" event={"ID":"1513698d-bd5c-41e8-92e3-cd9ee11ad479","Type":"ContainerDied","Data":"9d51de49bcf89ec563eabbe282e0f7116aaeae6a4093531e584b68bdc46b22a9"} Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.004582 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.131634 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-credential-keys\") pod \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.132180 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-fernet-keys\") pod \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.132202 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-combined-ca-bundle\") pod \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.132245 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm9vz\" (UniqueName: \"kubernetes.io/projected/1513698d-bd5c-41e8-92e3-cd9ee11ad479-kube-api-access-hm9vz\") pod \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.132273 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-config-data\") pod \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.132333 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-scripts\") pod \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\" (UID: \"1513698d-bd5c-41e8-92e3-cd9ee11ad479\") " Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.138375 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1513698d-bd5c-41e8-92e3-cd9ee11ad479" (UID: "1513698d-bd5c-41e8-92e3-cd9ee11ad479"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.138939 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-scripts" (OuterVolumeSpecName: "scripts") pod "1513698d-bd5c-41e8-92e3-cd9ee11ad479" (UID: "1513698d-bd5c-41e8-92e3-cd9ee11ad479"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.143251 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1513698d-bd5c-41e8-92e3-cd9ee11ad479" (UID: "1513698d-bd5c-41e8-92e3-cd9ee11ad479"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.149900 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1513698d-bd5c-41e8-92e3-cd9ee11ad479-kube-api-access-hm9vz" (OuterVolumeSpecName: "kube-api-access-hm9vz") pod "1513698d-bd5c-41e8-92e3-cd9ee11ad479" (UID: "1513698d-bd5c-41e8-92e3-cd9ee11ad479"). InnerVolumeSpecName "kube-api-access-hm9vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.159069 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-config-data" (OuterVolumeSpecName: "config-data") pod "1513698d-bd5c-41e8-92e3-cd9ee11ad479" (UID: "1513698d-bd5c-41e8-92e3-cd9ee11ad479"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.159457 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1513698d-bd5c-41e8-92e3-cd9ee11ad479" (UID: "1513698d-bd5c-41e8-92e3-cd9ee11ad479"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.234978 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.235020 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.235031 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.235063 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm9vz\" (UniqueName: \"kubernetes.io/projected/1513698d-bd5c-41e8-92e3-cd9ee11ad479-kube-api-access-hm9vz\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.235077 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.235087 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1513698d-bd5c-41e8-92e3-cd9ee11ad479-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.635943 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wmx87" event={"ID":"1513698d-bd5c-41e8-92e3-cd9ee11ad479","Type":"ContainerDied","Data":"de1a3526623ad21ed1d5dca6e4b5f700c83e111d532b492f7cd2ff31b24720d2"} Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.636009 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1a3526623ad21ed1d5dca6e4b5f700c83e111d532b492f7cd2ff31b24720d2" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.636032 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wmx87" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.728726 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wmx87"] Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.738387 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wmx87"] Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.806386 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-spqsj"] Dec 03 18:27:28 crc kubenswrapper[4758]: E1203 18:27:28.806727 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1513698d-bd5c-41e8-92e3-cd9ee11ad479" containerName="keystone-bootstrap" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.806745 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1513698d-bd5c-41e8-92e3-cd9ee11ad479" containerName="keystone-bootstrap" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.806912 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1513698d-bd5c-41e8-92e3-cd9ee11ad479" containerName="keystone-bootstrap" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.807504 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.810072 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.810453 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.811503 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.811734 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.816034 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-spqsj"] Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.862636 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gp4jx" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.967743 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-config-data\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.967968 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-combined-ca-bundle\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.968131 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-credential-keys\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.968221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-scripts\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.968367 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvpp8\" (UniqueName: \"kubernetes.io/projected/53831e41-c824-4d29-8d23-1c2db2e44ca9-kube-api-access-rvpp8\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:28 crc kubenswrapper[4758]: I1203 18:27:28.968462 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-fernet-keys\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.069888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvpp8\" (UniqueName: \"kubernetes.io/projected/53831e41-c824-4d29-8d23-1c2db2e44ca9-kube-api-access-rvpp8\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.070347 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-fernet-keys\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.071110 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-config-data\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.071156 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-combined-ca-bundle\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.071206 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-credential-keys\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.071246 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-scripts\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.075441 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-credential-keys\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.075461 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-combined-ca-bundle\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.075820 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-scripts\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.076652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-fernet-keys\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.076699 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-config-data\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.085219 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvpp8\" (UniqueName: \"kubernetes.io/projected/53831e41-c824-4d29-8d23-1c2db2e44ca9-kube-api-access-rvpp8\") pod \"keystone-bootstrap-spqsj\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.131190 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1513698d-bd5c-41e8-92e3-cd9ee11ad479" path="/var/lib/kubelet/pods/1513698d-bd5c-41e8-92e3-cd9ee11ad479/volumes" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.173221 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.620563 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-spqsj"] Dec 03 18:27:29 crc kubenswrapper[4758]: W1203 18:27:29.627422 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53831e41_c824_4d29_8d23_1c2db2e44ca9.slice/crio-b93387d4f6f5e786bb7e0390f80c6b7718dbd6d3ece3b374d43bb71ed626df1d WatchSource:0}: Error finding container b93387d4f6f5e786bb7e0390f80c6b7718dbd6d3ece3b374d43bb71ed626df1d: Status 404 returned error can't find the container with id b93387d4f6f5e786bb7e0390f80c6b7718dbd6d3ece3b374d43bb71ed626df1d Dec 03 18:27:29 crc kubenswrapper[4758]: I1203 18:27:29.644640 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spqsj" event={"ID":"53831e41-c824-4d29-8d23-1c2db2e44ca9","Type":"ContainerStarted","Data":"b93387d4f6f5e786bb7e0390f80c6b7718dbd6d3ece3b374d43bb71ed626df1d"} Dec 03 18:27:30 crc kubenswrapper[4758]: I1203 18:27:30.658238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spqsj" event={"ID":"53831e41-c824-4d29-8d23-1c2db2e44ca9","Type":"ContainerStarted","Data":"353f25e9e0d987349abbc108e38be48376f64dac9ce271fe217f4bca1aee6390"} Dec 03 18:27:30 crc kubenswrapper[4758]: I1203 18:27:30.684057 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-spqsj" podStartSLOduration=2.68403063 podStartE2EDuration="2.68403063s" podCreationTimestamp="2025-12-03 18:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:27:30.676086565 +0000 UTC m=+5505.877463466" watchObservedRunningTime="2025-12-03 18:27:30.68403063 +0000 UTC m=+5505.885407531" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.133876 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.218948 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86c4f5d6c9-xzm5z"] Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.219326 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" podUID="0d56e470-87c2-482f-858b-540fb535086e" containerName="dnsmasq-dns" containerID="cri-o://ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be" gracePeriod=10 Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.669859 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.673662 4758 generic.go:334] "Generic (PLEG): container finished" podID="53831e41-c824-4d29-8d23-1c2db2e44ca9" containerID="353f25e9e0d987349abbc108e38be48376f64dac9ce271fe217f4bca1aee6390" exitCode=0 Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.673708 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spqsj" event={"ID":"53831e41-c824-4d29-8d23-1c2db2e44ca9","Type":"ContainerDied","Data":"353f25e9e0d987349abbc108e38be48376f64dac9ce271fe217f4bca1aee6390"} Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.675951 4758 generic.go:334] "Generic (PLEG): container finished" podID="0d56e470-87c2-482f-858b-540fb535086e" containerID="ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be" exitCode=0 Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.675982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" event={"ID":"0d56e470-87c2-482f-858b-540fb535086e","Type":"ContainerDied","Data":"ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be"} Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.676007 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" event={"ID":"0d56e470-87c2-482f-858b-540fb535086e","Type":"ContainerDied","Data":"5b85f5131f9aa55d102bbff77d488dc802ac8088382f5f9c15e97026bc6cd705"} Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.676019 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c4f5d6c9-xzm5z" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.676023 4758 scope.go:117] "RemoveContainer" containerID="ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.702966 4758 scope.go:117] "RemoveContainer" containerID="a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.724933 4758 scope.go:117] "RemoveContainer" containerID="ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be" Dec 03 18:27:32 crc kubenswrapper[4758]: E1203 18:27:32.725465 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be\": container with ID starting with ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be not found: ID does not exist" containerID="ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.725542 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be"} err="failed to get container status \"ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be\": rpc error: code = NotFound desc = could not find container \"ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be\": container with ID starting with ee5ffd53ebf5c278bbc2f8f1bd12f9ad346a271bbcaf5a032814f35751c708be not found: ID does not exist" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.725600 4758 scope.go:117] "RemoveContainer" containerID="a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb" Dec 03 18:27:32 crc kubenswrapper[4758]: E1203 18:27:32.725998 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb\": container with ID starting with a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb not found: ID does not exist" containerID="a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.726050 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb"} err="failed to get container status \"a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb\": rpc error: code = NotFound desc = could not find container \"a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb\": container with ID starting with a72fcc33293eff4db9423da3f94e71c498bfa501ac28ea781a67c9f39a161aeb not found: ID does not exist" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.834097 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-dns-svc\") pod \"0d56e470-87c2-482f-858b-540fb535086e\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.835442 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-sb\") pod \"0d56e470-87c2-482f-858b-540fb535086e\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.835538 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-nb\") pod \"0d56e470-87c2-482f-858b-540fb535086e\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.835588 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-config\") pod \"0d56e470-87c2-482f-858b-540fb535086e\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.835608 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brtfz\" (UniqueName: \"kubernetes.io/projected/0d56e470-87c2-482f-858b-540fb535086e-kube-api-access-brtfz\") pod \"0d56e470-87c2-482f-858b-540fb535086e\" (UID: \"0d56e470-87c2-482f-858b-540fb535086e\") " Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.845012 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d56e470-87c2-482f-858b-540fb535086e-kube-api-access-brtfz" (OuterVolumeSpecName: "kube-api-access-brtfz") pod "0d56e470-87c2-482f-858b-540fb535086e" (UID: "0d56e470-87c2-482f-858b-540fb535086e"). InnerVolumeSpecName "kube-api-access-brtfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.873190 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0d56e470-87c2-482f-858b-540fb535086e" (UID: "0d56e470-87c2-482f-858b-540fb535086e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.873749 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d56e470-87c2-482f-858b-540fb535086e" (UID: "0d56e470-87c2-482f-858b-540fb535086e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.877179 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-config" (OuterVolumeSpecName: "config") pod "0d56e470-87c2-482f-858b-540fb535086e" (UID: "0d56e470-87c2-482f-858b-540fb535086e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.884496 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d56e470-87c2-482f-858b-540fb535086e" (UID: "0d56e470-87c2-482f-858b-540fb535086e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.937222 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.937260 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.937269 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brtfz\" (UniqueName: \"kubernetes.io/projected/0d56e470-87c2-482f-858b-540fb535086e-kube-api-access-brtfz\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.937279 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:32 crc kubenswrapper[4758]: I1203 18:27:32.937288 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d56e470-87c2-482f-858b-540fb535086e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:33 crc kubenswrapper[4758]: I1203 18:27:33.020481 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86c4f5d6c9-xzm5z"] Dec 03 18:27:33 crc kubenswrapper[4758]: I1203 18:27:33.029119 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86c4f5d6c9-xzm5z"] Dec 03 18:27:33 crc kubenswrapper[4758]: I1203 18:27:33.124316 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d56e470-87c2-482f-858b-540fb535086e" path="/var/lib/kubelet/pods/0d56e470-87c2-482f-858b-540fb535086e/volumes" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.049227 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.155502 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-combined-ca-bundle\") pod \"53831e41-c824-4d29-8d23-1c2db2e44ca9\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.155612 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-credential-keys\") pod \"53831e41-c824-4d29-8d23-1c2db2e44ca9\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.155718 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvpp8\" (UniqueName: \"kubernetes.io/projected/53831e41-c824-4d29-8d23-1c2db2e44ca9-kube-api-access-rvpp8\") pod \"53831e41-c824-4d29-8d23-1c2db2e44ca9\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.155754 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-fernet-keys\") pod \"53831e41-c824-4d29-8d23-1c2db2e44ca9\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.155867 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-config-data\") pod \"53831e41-c824-4d29-8d23-1c2db2e44ca9\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.155930 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-scripts\") pod \"53831e41-c824-4d29-8d23-1c2db2e44ca9\" (UID: \"53831e41-c824-4d29-8d23-1c2db2e44ca9\") " Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.159033 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "53831e41-c824-4d29-8d23-1c2db2e44ca9" (UID: "53831e41-c824-4d29-8d23-1c2db2e44ca9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.159648 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-scripts" (OuterVolumeSpecName: "scripts") pod "53831e41-c824-4d29-8d23-1c2db2e44ca9" (UID: "53831e41-c824-4d29-8d23-1c2db2e44ca9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.160512 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53831e41-c824-4d29-8d23-1c2db2e44ca9-kube-api-access-rvpp8" (OuterVolumeSpecName: "kube-api-access-rvpp8") pod "53831e41-c824-4d29-8d23-1c2db2e44ca9" (UID: "53831e41-c824-4d29-8d23-1c2db2e44ca9"). InnerVolumeSpecName "kube-api-access-rvpp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.161392 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "53831e41-c824-4d29-8d23-1c2db2e44ca9" (UID: "53831e41-c824-4d29-8d23-1c2db2e44ca9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.188037 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53831e41-c824-4d29-8d23-1c2db2e44ca9" (UID: "53831e41-c824-4d29-8d23-1c2db2e44ca9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.190350 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-config-data" (OuterVolumeSpecName: "config-data") pod "53831e41-c824-4d29-8d23-1c2db2e44ca9" (UID: "53831e41-c824-4d29-8d23-1c2db2e44ca9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.258299 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.258365 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.258394 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.258425 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.258455 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvpp8\" (UniqueName: \"kubernetes.io/projected/53831e41-c824-4d29-8d23-1c2db2e44ca9-kube-api-access-rvpp8\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.258481 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53831e41-c824-4d29-8d23-1c2db2e44ca9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.701160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spqsj" event={"ID":"53831e41-c824-4d29-8d23-1c2db2e44ca9","Type":"ContainerDied","Data":"b93387d4f6f5e786bb7e0390f80c6b7718dbd6d3ece3b374d43bb71ed626df1d"} Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.701203 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b93387d4f6f5e786bb7e0390f80c6b7718dbd6d3ece3b374d43bb71ed626df1d" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.701263 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spqsj" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.825932 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7669c5b48c-b2qhl"] Dec 03 18:27:34 crc kubenswrapper[4758]: E1203 18:27:34.826337 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d56e470-87c2-482f-858b-540fb535086e" containerName="init" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.826359 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d56e470-87c2-482f-858b-540fb535086e" containerName="init" Dec 03 18:27:34 crc kubenswrapper[4758]: E1203 18:27:34.826390 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53831e41-c824-4d29-8d23-1c2db2e44ca9" containerName="keystone-bootstrap" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.826399 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="53831e41-c824-4d29-8d23-1c2db2e44ca9" containerName="keystone-bootstrap" Dec 03 18:27:34 crc kubenswrapper[4758]: E1203 18:27:34.826416 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d56e470-87c2-482f-858b-540fb535086e" containerName="dnsmasq-dns" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.826426 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d56e470-87c2-482f-858b-540fb535086e" containerName="dnsmasq-dns" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.826632 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d56e470-87c2-482f-858b-540fb535086e" containerName="dnsmasq-dns" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.826671 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="53831e41-c824-4d29-8d23-1c2db2e44ca9" containerName="keystone-bootstrap" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.827530 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.830449 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gp4jx" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.830463 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.830463 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.831438 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.848975 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7669c5b48c-b2qhl"] Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.971751 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzm8k\" (UniqueName: \"kubernetes.io/projected/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-kube-api-access-vzm8k\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.971823 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-config-data\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.971849 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-fernet-keys\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.972378 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-scripts\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.972449 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-combined-ca-bundle\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:34 crc kubenswrapper[4758]: I1203 18:27:34.972498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-credential-keys\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.074490 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-scripts\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.074551 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-combined-ca-bundle\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.074588 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-credential-keys\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.074646 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzm8k\" (UniqueName: \"kubernetes.io/projected/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-kube-api-access-vzm8k\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.074724 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-config-data\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.074753 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-fernet-keys\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.079428 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-credential-keys\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.081433 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-fernet-keys\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.082971 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-scripts\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.086063 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-config-data\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.087233 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-combined-ca-bundle\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.111195 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzm8k\" (UniqueName: \"kubernetes.io/projected/7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f-kube-api-access-vzm8k\") pod \"keystone-7669c5b48c-b2qhl\" (UID: \"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f\") " pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.149105 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.608222 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7669c5b48c-b2qhl"] Dec 03 18:27:35 crc kubenswrapper[4758]: I1203 18:27:35.711696 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7669c5b48c-b2qhl" event={"ID":"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f","Type":"ContainerStarted","Data":"3f812bdd30b6606ee6ff9af2c35dd99bb301555b075e909f419e7a34b4e25c0b"} Dec 03 18:27:36 crc kubenswrapper[4758]: I1203 18:27:36.725658 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7669c5b48c-b2qhl" event={"ID":"7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f","Type":"ContainerStarted","Data":"f34b6ef6d1b1b2e6d97952b08280efd426fbcf5f51ad770b59f9e9e369d11d95"} Dec 03 18:27:36 crc kubenswrapper[4758]: I1203 18:27:36.725938 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:27:36 crc kubenswrapper[4758]: I1203 18:27:36.750590 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7669c5b48c-b2qhl" podStartSLOduration=2.7505682179999997 podStartE2EDuration="2.750568218s" podCreationTimestamp="2025-12-03 18:27:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:27:36.746004033 +0000 UTC m=+5511.947380924" watchObservedRunningTime="2025-12-03 18:27:36.750568218 +0000 UTC m=+5511.951945089" Dec 03 18:28:06 crc kubenswrapper[4758]: I1203 18:28:06.732335 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7669c5b48c-b2qhl" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.854694 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.856487 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.858295 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-48msh" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.859340 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.864004 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.866934 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.867351 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config-secret\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.867483 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.867590 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98ct4\" (UniqueName: \"kubernetes.io/projected/90b85a0f-5936-4bab-a550-83934e007e9b-kube-api-access-98ct4\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.969397 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config-secret\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.969469 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.969506 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98ct4\" (UniqueName: \"kubernetes.io/projected/90b85a0f-5936-4bab-a550-83934e007e9b-kube-api-access-98ct4\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.970898 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.976611 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config-secret\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:10 crc kubenswrapper[4758]: I1203 18:28:10.983965 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98ct4\" (UniqueName: \"kubernetes.io/projected/90b85a0f-5936-4bab-a550-83934e007e9b-kube-api-access-98ct4\") pod \"openstackclient\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " pod="openstack/openstackclient" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.084056 4758 scope.go:117] "RemoveContainer" containerID="170a6f6514c8e4c51d323009bdfbde47cc633aceca2bc468502459ac0232341f" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.106032 4758 scope.go:117] "RemoveContainer" containerID="1a0e51cfde4ed5e4594e8cd09e944ec56b1803e593417d1381a7c40d42a3247d" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.157447 4758 scope.go:117] "RemoveContainer" containerID="20c83482882c4646603895b967a564b19d61051fcfb9475c32608947171d8448" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.191192 4758 scope.go:117] "RemoveContainer" containerID="5b9a908934fa2673abce87a39335c304f7f795214c900c7ef84a5a2ec6925e8f" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.191444 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.394544 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.394618 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:28:11 crc kubenswrapper[4758]: I1203 18:28:11.619566 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 18:28:11 crc kubenswrapper[4758]: W1203 18:28:11.622468 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90b85a0f_5936_4bab_a550_83934e007e9b.slice/crio-d199e7059f564fc98c173560096939997337d4e444611cfe2f14e1c9b6bad54e WatchSource:0}: Error finding container d199e7059f564fc98c173560096939997337d4e444611cfe2f14e1c9b6bad54e: Status 404 returned error can't find the container with id d199e7059f564fc98c173560096939997337d4e444611cfe2f14e1c9b6bad54e Dec 03 18:28:12 crc kubenswrapper[4758]: I1203 18:28:12.091553 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"90b85a0f-5936-4bab-a550-83934e007e9b","Type":"ContainerStarted","Data":"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6"} Dec 03 18:28:12 crc kubenswrapper[4758]: I1203 18:28:12.091887 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"90b85a0f-5936-4bab-a550-83934e007e9b","Type":"ContainerStarted","Data":"d199e7059f564fc98c173560096939997337d4e444611cfe2f14e1c9b6bad54e"} Dec 03 18:28:12 crc kubenswrapper[4758]: I1203 18:28:12.108924 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.108904058 podStartE2EDuration="2.108904058s" podCreationTimestamp="2025-12-03 18:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:28:12.105621501 +0000 UTC m=+5547.306998382" watchObservedRunningTime="2025-12-03 18:28:12.108904058 +0000 UTC m=+5547.310280939" Dec 03 18:28:41 crc kubenswrapper[4758]: I1203 18:28:41.396178 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:28:41 crc kubenswrapper[4758]: I1203 18:28:41.397717 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.395219 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.395851 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.395942 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.397471 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee02f3819ed4276671e01c5782910e87da9be6dd0555ff89c3f0a62d67c490da"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.397577 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://ee02f3819ed4276671e01c5782910e87da9be6dd0555ff89c3f0a62d67c490da" gracePeriod=600 Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.652097 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="ee02f3819ed4276671e01c5782910e87da9be6dd0555ff89c3f0a62d67c490da" exitCode=0 Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.652155 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"ee02f3819ed4276671e01c5782910e87da9be6dd0555ff89c3f0a62d67c490da"} Dec 03 18:29:11 crc kubenswrapper[4758]: I1203 18:29:11.652206 4758 scope.go:117] "RemoveContainer" containerID="fa08fe454fc2fd1472e152881ecb5b188eb6d6fea96c54a68f57f71b1ef5e176" Dec 03 18:29:12 crc kubenswrapper[4758]: I1203 18:29:12.662840 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3"} Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.141995 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp"] Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.143545 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.145146 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.145400 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.155729 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp"] Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.333652 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5h8n\" (UniqueName: \"kubernetes.io/projected/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-kube-api-access-q5h8n\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.333748 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-config-volume\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.334551 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-secret-volume\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.435860 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-secret-volume\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.436327 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5h8n\" (UniqueName: \"kubernetes.io/projected/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-kube-api-access-q5h8n\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.436373 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-config-volume\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.437491 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-config-volume\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.449011 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-secret-volume\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.461894 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5h8n\" (UniqueName: \"kubernetes.io/projected/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-kube-api-access-q5h8n\") pod \"collect-profiles-29413110-h6czp\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.467082 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:00 crc kubenswrapper[4758]: I1203 18:30:00.892990 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp"] Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.347398 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" event={"ID":"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f","Type":"ContainerStarted","Data":"cb1afbd76be525c4ef275e5759cce59a95c225260106d9452636009eed098033"} Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.580254 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5a20-account-create-update-wd9nn"] Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.581506 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.585511 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.586007 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-kf7sz"] Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.587225 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.592225 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a20-account-create-update-wd9nn"] Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.599202 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kf7sz"] Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.656827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/804c639b-8c29-4d6c-864a-77715bf8b004-operator-scripts\") pod \"barbican-5a20-account-create-update-wd9nn\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.656919 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm8nb\" (UniqueName: \"kubernetes.io/projected/4ddf7b6b-9452-49ba-b875-6caf9392059f-kube-api-access-hm8nb\") pod \"barbican-db-create-kf7sz\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.656974 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ddf7b6b-9452-49ba-b875-6caf9392059f-operator-scripts\") pod \"barbican-db-create-kf7sz\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.657012 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vddv9\" (UniqueName: \"kubernetes.io/projected/804c639b-8c29-4d6c-864a-77715bf8b004-kube-api-access-vddv9\") pod \"barbican-5a20-account-create-update-wd9nn\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.757770 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm8nb\" (UniqueName: \"kubernetes.io/projected/4ddf7b6b-9452-49ba-b875-6caf9392059f-kube-api-access-hm8nb\") pod \"barbican-db-create-kf7sz\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.757843 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ddf7b6b-9452-49ba-b875-6caf9392059f-operator-scripts\") pod \"barbican-db-create-kf7sz\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.757879 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vddv9\" (UniqueName: \"kubernetes.io/projected/804c639b-8c29-4d6c-864a-77715bf8b004-kube-api-access-vddv9\") pod \"barbican-5a20-account-create-update-wd9nn\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.757956 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/804c639b-8c29-4d6c-864a-77715bf8b004-operator-scripts\") pod \"barbican-5a20-account-create-update-wd9nn\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.759095 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ddf7b6b-9452-49ba-b875-6caf9392059f-operator-scripts\") pod \"barbican-db-create-kf7sz\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.759112 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/804c639b-8c29-4d6c-864a-77715bf8b004-operator-scripts\") pod \"barbican-5a20-account-create-update-wd9nn\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.778577 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vddv9\" (UniqueName: \"kubernetes.io/projected/804c639b-8c29-4d6c-864a-77715bf8b004-kube-api-access-vddv9\") pod \"barbican-5a20-account-create-update-wd9nn\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.790460 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm8nb\" (UniqueName: \"kubernetes.io/projected/4ddf7b6b-9452-49ba-b875-6caf9392059f-kube-api-access-hm8nb\") pod \"barbican-db-create-kf7sz\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.901857 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:01 crc kubenswrapper[4758]: I1203 18:30:01.911307 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:02 crc kubenswrapper[4758]: I1203 18:30:02.363190 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" containerID="f8071202aca079ccd91262d1c98751537a40a01ad2c74d5b20373ff7c5bf4d86" exitCode=0 Dec 03 18:30:02 crc kubenswrapper[4758]: I1203 18:30:02.363440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" event={"ID":"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f","Type":"ContainerDied","Data":"f8071202aca079ccd91262d1c98751537a40a01ad2c74d5b20373ff7c5bf4d86"} Dec 03 18:30:02 crc kubenswrapper[4758]: I1203 18:30:02.456942 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a20-account-create-update-wd9nn"] Dec 03 18:30:02 crc kubenswrapper[4758]: I1203 18:30:02.567307 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kf7sz"] Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.377139 4758 generic.go:334] "Generic (PLEG): container finished" podID="4ddf7b6b-9452-49ba-b875-6caf9392059f" containerID="10f5969c64aa23c105137b7c839a8316a6ac45aa1709a9d7cf6015009416420a" exitCode=0 Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.377225 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kf7sz" event={"ID":"4ddf7b6b-9452-49ba-b875-6caf9392059f","Type":"ContainerDied","Data":"10f5969c64aa23c105137b7c839a8316a6ac45aa1709a9d7cf6015009416420a"} Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.377319 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kf7sz" event={"ID":"4ddf7b6b-9452-49ba-b875-6caf9392059f","Type":"ContainerStarted","Data":"c8bec54f720ed9a19c1d2f74aab3530bdde3a3652aa29985bbb895a228fe392c"} Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.379272 4758 generic.go:334] "Generic (PLEG): container finished" podID="804c639b-8c29-4d6c-864a-77715bf8b004" containerID="90dfacb6f481088b94e0cee74f4e22f3ecbc31330fd7af7f2df0c004833529f3" exitCode=0 Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.379348 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a20-account-create-update-wd9nn" event={"ID":"804c639b-8c29-4d6c-864a-77715bf8b004","Type":"ContainerDied","Data":"90dfacb6f481088b94e0cee74f4e22f3ecbc31330fd7af7f2df0c004833529f3"} Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.379424 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a20-account-create-update-wd9nn" event={"ID":"804c639b-8c29-4d6c-864a-77715bf8b004","Type":"ContainerStarted","Data":"73f1a2db021812e9f86412cf25b2edf35a435fefae2dc58c56367494c5d67a4f"} Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.726930 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.907037 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5h8n\" (UniqueName: \"kubernetes.io/projected/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-kube-api-access-q5h8n\") pod \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.907088 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-config-volume\") pod \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.907154 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-secret-volume\") pod \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\" (UID: \"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f\") " Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.908040 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-config-volume" (OuterVolumeSpecName: "config-volume") pod "7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" (UID: "7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.915628 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-kube-api-access-q5h8n" (OuterVolumeSpecName: "kube-api-access-q5h8n") pod "7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" (UID: "7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f"). InnerVolumeSpecName "kube-api-access-q5h8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4758]: I1203 18:30:03.917814 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" (UID: "7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.009853 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.009892 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.009905 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5h8n\" (UniqueName: \"kubernetes.io/projected/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f-kube-api-access-q5h8n\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.391618 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" event={"ID":"7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f","Type":"ContainerDied","Data":"cb1afbd76be525c4ef275e5759cce59a95c225260106d9452636009eed098033"} Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.391672 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb1afbd76be525c4ef275e5759cce59a95c225260106d9452636009eed098033" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.394661 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.728135 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.739115 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.801891 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8"] Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.807655 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-5pnx8"] Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.924502 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ddf7b6b-9452-49ba-b875-6caf9392059f-operator-scripts\") pod \"4ddf7b6b-9452-49ba-b875-6caf9392059f\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.925299 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddf7b6b-9452-49ba-b875-6caf9392059f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ddf7b6b-9452-49ba-b875-6caf9392059f" (UID: "4ddf7b6b-9452-49ba-b875-6caf9392059f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.925444 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/804c639b-8c29-4d6c-864a-77715bf8b004-operator-scripts\") pod \"804c639b-8c29-4d6c-864a-77715bf8b004\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.925913 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/804c639b-8c29-4d6c-864a-77715bf8b004-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "804c639b-8c29-4d6c-864a-77715bf8b004" (UID: "804c639b-8c29-4d6c-864a-77715bf8b004"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.926047 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vddv9\" (UniqueName: \"kubernetes.io/projected/804c639b-8c29-4d6c-864a-77715bf8b004-kube-api-access-vddv9\") pod \"804c639b-8c29-4d6c-864a-77715bf8b004\" (UID: \"804c639b-8c29-4d6c-864a-77715bf8b004\") " Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.926248 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm8nb\" (UniqueName: \"kubernetes.io/projected/4ddf7b6b-9452-49ba-b875-6caf9392059f-kube-api-access-hm8nb\") pod \"4ddf7b6b-9452-49ba-b875-6caf9392059f\" (UID: \"4ddf7b6b-9452-49ba-b875-6caf9392059f\") " Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.927054 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ddf7b6b-9452-49ba-b875-6caf9392059f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.927256 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/804c639b-8c29-4d6c-864a-77715bf8b004-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.930458 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddf7b6b-9452-49ba-b875-6caf9392059f-kube-api-access-hm8nb" (OuterVolumeSpecName: "kube-api-access-hm8nb") pod "4ddf7b6b-9452-49ba-b875-6caf9392059f" (UID: "4ddf7b6b-9452-49ba-b875-6caf9392059f"). InnerVolumeSpecName "kube-api-access-hm8nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:04 crc kubenswrapper[4758]: I1203 18:30:04.930843 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/804c639b-8c29-4d6c-864a-77715bf8b004-kube-api-access-vddv9" (OuterVolumeSpecName: "kube-api-access-vddv9") pod "804c639b-8c29-4d6c-864a-77715bf8b004" (UID: "804c639b-8c29-4d6c-864a-77715bf8b004"). InnerVolumeSpecName "kube-api-access-vddv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.028888 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vddv9\" (UniqueName: \"kubernetes.io/projected/804c639b-8c29-4d6c-864a-77715bf8b004-kube-api-access-vddv9\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.028922 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm8nb\" (UniqueName: \"kubernetes.io/projected/4ddf7b6b-9452-49ba-b875-6caf9392059f-kube-api-access-hm8nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.123795 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f265e5-e02b-4b73-880b-d1993e88a87f" path="/var/lib/kubelet/pods/06f265e5-e02b-4b73-880b-d1993e88a87f/volumes" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.404062 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a20-account-create-update-wd9nn" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.404569 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a20-account-create-update-wd9nn" event={"ID":"804c639b-8c29-4d6c-864a-77715bf8b004","Type":"ContainerDied","Data":"73f1a2db021812e9f86412cf25b2edf35a435fefae2dc58c56367494c5d67a4f"} Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.405799 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73f1a2db021812e9f86412cf25b2edf35a435fefae2dc58c56367494c5d67a4f" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.406617 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kf7sz" event={"ID":"4ddf7b6b-9452-49ba-b875-6caf9392059f","Type":"ContainerDied","Data":"c8bec54f720ed9a19c1d2f74aab3530bdde3a3652aa29985bbb895a228fe392c"} Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.406669 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8bec54f720ed9a19c1d2f74aab3530bdde3a3652aa29985bbb895a228fe392c" Dec 03 18:30:05 crc kubenswrapper[4758]: I1203 18:30:05.406634 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kf7sz" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.859109 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-v9vjh"] Dec 03 18:30:06 crc kubenswrapper[4758]: E1203 18:30:06.859996 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddf7b6b-9452-49ba-b875-6caf9392059f" containerName="mariadb-database-create" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860010 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddf7b6b-9452-49ba-b875-6caf9392059f" containerName="mariadb-database-create" Dec 03 18:30:06 crc kubenswrapper[4758]: E1203 18:30:06.860032 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="804c639b-8c29-4d6c-864a-77715bf8b004" containerName="mariadb-account-create-update" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860040 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="804c639b-8c29-4d6c-864a-77715bf8b004" containerName="mariadb-account-create-update" Dec 03 18:30:06 crc kubenswrapper[4758]: E1203 18:30:06.860068 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" containerName="collect-profiles" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860074 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" containerName="collect-profiles" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860247 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="804c639b-8c29-4d6c-864a-77715bf8b004" containerName="mariadb-account-create-update" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860262 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddf7b6b-9452-49ba-b875-6caf9392059f" containerName="mariadb-database-create" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860278 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" containerName="collect-profiles" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.860895 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.864413 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hbgl6" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.864654 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.876291 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-v9vjh"] Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.958023 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-db-sync-config-data\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.959710 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2x2t\" (UniqueName: \"kubernetes.io/projected/9547728c-ddd1-4caa-b665-b584c15304ac-kube-api-access-k2x2t\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:06 crc kubenswrapper[4758]: I1203 18:30:06.959833 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-combined-ca-bundle\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.061720 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-db-sync-config-data\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.061794 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2x2t\" (UniqueName: \"kubernetes.io/projected/9547728c-ddd1-4caa-b665-b584c15304ac-kube-api-access-k2x2t\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.061817 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-combined-ca-bundle\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.067494 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-combined-ca-bundle\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.074236 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-db-sync-config-data\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.083516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2x2t\" (UniqueName: \"kubernetes.io/projected/9547728c-ddd1-4caa-b665-b584c15304ac-kube-api-access-k2x2t\") pod \"barbican-db-sync-v9vjh\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.181747 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:07 crc kubenswrapper[4758]: I1203 18:30:07.706052 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-v9vjh"] Dec 03 18:30:08 crc kubenswrapper[4758]: I1203 18:30:08.430971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v9vjh" event={"ID":"9547728c-ddd1-4caa-b665-b584c15304ac","Type":"ContainerStarted","Data":"2b240713afa4d648a548d6384c263cfa32026d4a07dc55905181b8b9ab3abdc9"} Dec 03 18:30:09 crc kubenswrapper[4758]: I1203 18:30:09.439047 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v9vjh" event={"ID":"9547728c-ddd1-4caa-b665-b584c15304ac","Type":"ContainerStarted","Data":"1f9beb59e1ebac57a7b7f152c1626cc803ccccf03c7475d421471baaa2aa1c48"} Dec 03 18:30:09 crc kubenswrapper[4758]: I1203 18:30:09.463045 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-v9vjh" podStartSLOduration=3.463020401 podStartE2EDuration="3.463020401s" podCreationTimestamp="2025-12-03 18:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:09.453375816 +0000 UTC m=+5664.654752677" watchObservedRunningTime="2025-12-03 18:30:09.463020401 +0000 UTC m=+5664.664397272" Dec 03 18:30:10 crc kubenswrapper[4758]: I1203 18:30:10.476931 4758 generic.go:334] "Generic (PLEG): container finished" podID="9547728c-ddd1-4caa-b665-b584c15304ac" containerID="1f9beb59e1ebac57a7b7f152c1626cc803ccccf03c7475d421471baaa2aa1c48" exitCode=0 Dec 03 18:30:10 crc kubenswrapper[4758]: I1203 18:30:10.476976 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v9vjh" event={"ID":"9547728c-ddd1-4caa-b665-b584c15304ac","Type":"ContainerDied","Data":"1f9beb59e1ebac57a7b7f152c1626cc803ccccf03c7475d421471baaa2aa1c48"} Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.324937 4758 scope.go:117] "RemoveContainer" containerID="9b2289fc2ae3834503c845da9abb3a38e925e6829a3dee6f67bef5a7b8aaf298" Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.786223 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.896638 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2x2t\" (UniqueName: \"kubernetes.io/projected/9547728c-ddd1-4caa-b665-b584c15304ac-kube-api-access-k2x2t\") pod \"9547728c-ddd1-4caa-b665-b584c15304ac\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.896790 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-db-sync-config-data\") pod \"9547728c-ddd1-4caa-b665-b584c15304ac\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.897101 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-combined-ca-bundle\") pod \"9547728c-ddd1-4caa-b665-b584c15304ac\" (UID: \"9547728c-ddd1-4caa-b665-b584c15304ac\") " Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.905072 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9547728c-ddd1-4caa-b665-b584c15304ac" (UID: "9547728c-ddd1-4caa-b665-b584c15304ac"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.907917 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9547728c-ddd1-4caa-b665-b584c15304ac-kube-api-access-k2x2t" (OuterVolumeSpecName: "kube-api-access-k2x2t") pod "9547728c-ddd1-4caa-b665-b584c15304ac" (UID: "9547728c-ddd1-4caa-b665-b584c15304ac"). InnerVolumeSpecName "kube-api-access-k2x2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:11 crc kubenswrapper[4758]: I1203 18:30:11.921486 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9547728c-ddd1-4caa-b665-b584c15304ac" (UID: "9547728c-ddd1-4caa-b665-b584c15304ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.000062 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2x2t\" (UniqueName: \"kubernetes.io/projected/9547728c-ddd1-4caa-b665-b584c15304ac-kube-api-access-k2x2t\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.000121 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.000141 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9547728c-ddd1-4caa-b665-b584c15304ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.493845 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v9vjh" event={"ID":"9547728c-ddd1-4caa-b665-b584c15304ac","Type":"ContainerDied","Data":"2b240713afa4d648a548d6384c263cfa32026d4a07dc55905181b8b9ab3abdc9"} Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.493888 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b240713afa4d648a548d6384c263cfa32026d4a07dc55905181b8b9ab3abdc9" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.493953 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v9vjh" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.691154 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7f48c6d99f-5k4sg"] Dec 03 18:30:12 crc kubenswrapper[4758]: E1203 18:30:12.691483 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9547728c-ddd1-4caa-b665-b584c15304ac" containerName="barbican-db-sync" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.691500 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9547728c-ddd1-4caa-b665-b584c15304ac" containerName="barbican-db-sync" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.691800 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9547728c-ddd1-4caa-b665-b584c15304ac" containerName="barbican-db-sync" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.692611 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.698077 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.698150 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.698530 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hbgl6" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.726877 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7f48c6d99f-5k4sg"] Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.743601 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7b5dbc78b-xh7jg"] Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.746235 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.749899 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.772795 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b5dbc78b-xh7jg"] Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.816242 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-config-data-custom\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.816309 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-combined-ca-bundle\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.816342 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddhqc\" (UniqueName: \"kubernetes.io/projected/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-kube-api-access-ddhqc\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.816383 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-logs\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.816450 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-config-data\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.823957 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb"] Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.825550 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.865590 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb"] Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919303 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhxfn\" (UniqueName: \"kubernetes.io/projected/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-kube-api-access-lhxfn\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919347 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-config-data\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-config-data\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919398 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-config-data-custom\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919442 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-combined-ca-bundle\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919467 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddhqc\" (UniqueName: \"kubernetes.io/projected/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-kube-api-access-ddhqc\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-logs\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919520 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-config-data-custom\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919541 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-logs\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.919584 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-combined-ca-bundle\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.922115 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-logs\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.929825 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-config-data-custom\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.939026 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f468494d-9f2zm"] Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.939921 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddhqc\" (UniqueName: \"kubernetes.io/projected/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-kube-api-access-ddhqc\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.940935 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-combined-ca-bundle\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.941157 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.941872 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b740aeeb-6458-4a97-a7bc-a486f0ed2d97-config-data\") pod \"barbican-worker-7f48c6d99f-5k4sg\" (UID: \"b740aeeb-6458-4a97-a7bc-a486f0ed2d97\") " pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.945483 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 18:30:12 crc kubenswrapper[4758]: I1203 18:30:12.970099 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f468494d-9f2zm"] Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020463 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-config\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020866 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-logs\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzshv\" (UniqueName: \"kubernetes.io/projected/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-kube-api-access-tzshv\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020917 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-config-data-custom\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020938 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-dns-svc\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020965 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.020996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-combined-ca-bundle\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.021096 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.021266 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhxfn\" (UniqueName: \"kubernetes.io/projected/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-kube-api-access-lhxfn\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.021314 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-config-data\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.021557 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-logs\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.022185 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f48c6d99f-5k4sg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.027309 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-combined-ca-bundle\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.027450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-config-data-custom\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.028498 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-config-data\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.039238 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhxfn\" (UniqueName: \"kubernetes.io/projected/1f3421fd-2c4e-4ec0-9bca-345973cc7fb0-kube-api-access-lhxfn\") pod \"barbican-keystone-listener-7b5dbc78b-xh7jg\" (UID: \"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0\") " pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.070863 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123057 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-config\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123128 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-config-data\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123173 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-config-data-custom\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzshv\" (UniqueName: \"kubernetes.io/projected/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-kube-api-access-tzshv\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123254 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-dns-svc\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123290 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkgb2\" (UniqueName: \"kubernetes.io/projected/947fb70a-1772-4b0e-b6b2-eff63219feeb-kube-api-access-rkgb2\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123326 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123378 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-combined-ca-bundle\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123404 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123448 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947fb70a-1772-4b0e-b6b2-eff63219feeb-logs\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.123974 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-config\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.124476 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.124701 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-dns-svc\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.125016 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.146802 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzshv\" (UniqueName: \"kubernetes.io/projected/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-kube-api-access-tzshv\") pod \"dnsmasq-dns-6bf6b7dd8c-sr8fb\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.225346 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-config-data\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.225420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-config-data-custom\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.225491 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkgb2\" (UniqueName: \"kubernetes.io/projected/947fb70a-1772-4b0e-b6b2-eff63219feeb-kube-api-access-rkgb2\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.225539 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-combined-ca-bundle\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.225564 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947fb70a-1772-4b0e-b6b2-eff63219feeb-logs\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.225995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947fb70a-1772-4b0e-b6b2-eff63219feeb-logs\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.232385 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-combined-ca-bundle\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.232635 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-config-data\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.233185 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947fb70a-1772-4b0e-b6b2-eff63219feeb-config-data-custom\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.242649 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkgb2\" (UniqueName: \"kubernetes.io/projected/947fb70a-1772-4b0e-b6b2-eff63219feeb-kube-api-access-rkgb2\") pod \"barbican-api-5f468494d-9f2zm\" (UID: \"947fb70a-1772-4b0e-b6b2-eff63219feeb\") " pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.302873 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.417014 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b5dbc78b-xh7jg"] Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.441549 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.518816 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" event={"ID":"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0","Type":"ContainerStarted","Data":"6dd3b8853c3fe9a98584603b0d33d8ab2b08211a5a7f773814da005bde8261a2"} Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.559247 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7f48c6d99f-5k4sg"] Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.785280 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f468494d-9f2zm"] Dec 03 18:30:13 crc kubenswrapper[4758]: W1203 18:30:13.787386 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod947fb70a_1772_4b0e_b6b2_eff63219feeb.slice/crio-411bdbbb0a452835fa2da698f213a475964543da3c8ec3190964a353d7ee5e51 WatchSource:0}: Error finding container 411bdbbb0a452835fa2da698f213a475964543da3c8ec3190964a353d7ee5e51: Status 404 returned error can't find the container with id 411bdbbb0a452835fa2da698f213a475964543da3c8ec3190964a353d7ee5e51 Dec 03 18:30:13 crc kubenswrapper[4758]: I1203 18:30:13.911903 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb"] Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.531379 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" event={"ID":"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0","Type":"ContainerStarted","Data":"578db7e9a868c991383c6c821874882ae0173e5da94592096a6103b76e2df73d"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.531757 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" event={"ID":"1f3421fd-2c4e-4ec0-9bca-345973cc7fb0","Type":"ContainerStarted","Data":"7d9f67be35d0f39e1e1317838dc6fff39061e1e95a2f3a0804007678dc2ee33d"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.534321 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" event={"ID":"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1","Type":"ContainerStarted","Data":"da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.534370 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" event={"ID":"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1","Type":"ContainerStarted","Data":"df8bbb851a441d02e52554620c9fc6823ddab38fa66b911bb42372c14abf0840"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.542429 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f468494d-9f2zm" event={"ID":"947fb70a-1772-4b0e-b6b2-eff63219feeb","Type":"ContainerStarted","Data":"7200e542e3d74907c8f8fb4e32a81c34d7588f61713b9c30b8da32b15d5c6fe5"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.542473 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f468494d-9f2zm" event={"ID":"947fb70a-1772-4b0e-b6b2-eff63219feeb","Type":"ContainerStarted","Data":"00eba6d2d70286111719bc7532163489d038786400506a40bbdb2ea78d8025d1"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.542487 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f468494d-9f2zm" event={"ID":"947fb70a-1772-4b0e-b6b2-eff63219feeb","Type":"ContainerStarted","Data":"411bdbbb0a452835fa2da698f213a475964543da3c8ec3190964a353d7ee5e51"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.543221 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.543254 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.549484 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f48c6d99f-5k4sg" event={"ID":"b740aeeb-6458-4a97-a7bc-a486f0ed2d97","Type":"ContainerStarted","Data":"2278aef41cab20e236fb644cbdaa52fdadfd581d4b38680b6dc07c89e931ac60"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.549558 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f48c6d99f-5k4sg" event={"ID":"b740aeeb-6458-4a97-a7bc-a486f0ed2d97","Type":"ContainerStarted","Data":"5979e9c4cb41d262643b662be57ed69f66a0ae865523b60482dd3a1be086adba"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.549570 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f48c6d99f-5k4sg" event={"ID":"b740aeeb-6458-4a97-a7bc-a486f0ed2d97","Type":"ContainerStarted","Data":"e01b4998d2fb8abb9fce6b5933db6f00b273da55ff0744e3d4f8b9968b30c4a5"} Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.557669 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7b5dbc78b-xh7jg" podStartSLOduration=2.557652048 podStartE2EDuration="2.557652048s" podCreationTimestamp="2025-12-03 18:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:14.554518655 +0000 UTC m=+5669.755895516" watchObservedRunningTime="2025-12-03 18:30:14.557652048 +0000 UTC m=+5669.759028909" Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.610610 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7f48c6d99f-5k4sg" podStartSLOduration=2.610592626 podStartE2EDuration="2.610592626s" podCreationTimestamp="2025-12-03 18:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:14.583672899 +0000 UTC m=+5669.785049760" watchObservedRunningTime="2025-12-03 18:30:14.610592626 +0000 UTC m=+5669.811969487" Dec 03 18:30:14 crc kubenswrapper[4758]: I1203 18:30:14.635291 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f468494d-9f2zm" podStartSLOduration=2.635273608 podStartE2EDuration="2.635273608s" podCreationTimestamp="2025-12-03 18:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:14.629559839 +0000 UTC m=+5669.830936700" watchObservedRunningTime="2025-12-03 18:30:14.635273608 +0000 UTC m=+5669.836650459" Dec 03 18:30:15 crc kubenswrapper[4758]: I1203 18:30:15.560707 4758 generic.go:334] "Generic (PLEG): container finished" podID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerID="da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33" exitCode=0 Dec 03 18:30:15 crc kubenswrapper[4758]: I1203 18:30:15.560824 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" event={"ID":"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1","Type":"ContainerDied","Data":"da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33"} Dec 03 18:30:16 crc kubenswrapper[4758]: I1203 18:30:16.570767 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" event={"ID":"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1","Type":"ContainerStarted","Data":"2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3"} Dec 03 18:30:16 crc kubenswrapper[4758]: I1203 18:30:16.607417 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" podStartSLOduration=4.607393132 podStartE2EDuration="4.607393132s" podCreationTimestamp="2025-12-03 18:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:16.58608893 +0000 UTC m=+5671.787465791" watchObservedRunningTime="2025-12-03 18:30:16.607393132 +0000 UTC m=+5671.808769993" Dec 03 18:30:17 crc kubenswrapper[4758]: I1203 18:30:17.577362 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:19 crc kubenswrapper[4758]: I1203 18:30:19.842255 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:21 crc kubenswrapper[4758]: I1203 18:30:21.289542 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f468494d-9f2zm" Dec 03 18:30:23 crc kubenswrapper[4758]: I1203 18:30:23.442859 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:23 crc kubenswrapper[4758]: I1203 18:30:23.560504 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67674546dc-bp5pk"] Dec 03 18:30:23 crc kubenswrapper[4758]: I1203 18:30:23.560777 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerName="dnsmasq-dns" containerID="cri-o://48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08" gracePeriod=10 Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.513206 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.634473 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-sb\") pod \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.634935 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-nb\") pod \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.635002 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-config\") pod \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.635099 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz588\" (UniqueName: \"kubernetes.io/projected/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-kube-api-access-rz588\") pod \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.635133 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-dns-svc\") pod \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\" (UID: \"f9aad7c6-1569-4524-9f1e-c8240fd4ad51\") " Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.640930 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-kube-api-access-rz588" (OuterVolumeSpecName: "kube-api-access-rz588") pod "f9aad7c6-1569-4524-9f1e-c8240fd4ad51" (UID: "f9aad7c6-1569-4524-9f1e-c8240fd4ad51"). InnerVolumeSpecName "kube-api-access-rz588". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.650732 4758 generic.go:334] "Generic (PLEG): container finished" podID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerID="48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08" exitCode=0 Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.650785 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" event={"ID":"f9aad7c6-1569-4524-9f1e-c8240fd4ad51","Type":"ContainerDied","Data":"48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08"} Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.650815 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" event={"ID":"f9aad7c6-1569-4524-9f1e-c8240fd4ad51","Type":"ContainerDied","Data":"ed33543e41eebfd131e4f973991e9696f149d7c1d3376e6b9b6fd88bfcd1ff61"} Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.650835 4758 scope.go:117] "RemoveContainer" containerID="48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.650973 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67674546dc-bp5pk" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.675646 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-config" (OuterVolumeSpecName: "config") pod "f9aad7c6-1569-4524-9f1e-c8240fd4ad51" (UID: "f9aad7c6-1569-4524-9f1e-c8240fd4ad51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.683880 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f9aad7c6-1569-4524-9f1e-c8240fd4ad51" (UID: "f9aad7c6-1569-4524-9f1e-c8240fd4ad51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.683955 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f9aad7c6-1569-4524-9f1e-c8240fd4ad51" (UID: "f9aad7c6-1569-4524-9f1e-c8240fd4ad51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.695007 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f9aad7c6-1569-4524-9f1e-c8240fd4ad51" (UID: "f9aad7c6-1569-4524-9f1e-c8240fd4ad51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.737702 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz588\" (UniqueName: \"kubernetes.io/projected/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-kube-api-access-rz588\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.737829 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.737899 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.737956 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.738007 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9aad7c6-1569-4524-9f1e-c8240fd4ad51-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.754716 4758 scope.go:117] "RemoveContainer" containerID="52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.778573 4758 scope.go:117] "RemoveContainer" containerID="48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08" Dec 03 18:30:24 crc kubenswrapper[4758]: E1203 18:30:24.779221 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08\": container with ID starting with 48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08 not found: ID does not exist" containerID="48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.779376 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08"} err="failed to get container status \"48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08\": rpc error: code = NotFound desc = could not find container \"48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08\": container with ID starting with 48beabdaceac7f87d3c9bd578f9f1239ef99087c98f49c21dcb1c251c6e84f08 not found: ID does not exist" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.779508 4758 scope.go:117] "RemoveContainer" containerID="52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417" Dec 03 18:30:24 crc kubenswrapper[4758]: E1203 18:30:24.779942 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417\": container with ID starting with 52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417 not found: ID does not exist" containerID="52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.780095 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417"} err="failed to get container status \"52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417\": rpc error: code = NotFound desc = could not find container \"52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417\": container with ID starting with 52b66b3d1bd4e57709b6b48f91d3cd645e63e5e987d8ea433adfdb3ab8822417 not found: ID does not exist" Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.985299 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67674546dc-bp5pk"] Dec 03 18:30:24 crc kubenswrapper[4758]: I1203 18:30:24.993780 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67674546dc-bp5pk"] Dec 03 18:30:25 crc kubenswrapper[4758]: I1203 18:30:25.122867 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" path="/var/lib/kubelet/pods/f9aad7c6-1569-4524-9f1e-c8240fd4ad51/volumes" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.747114 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pfhjl"] Dec 03 18:30:33 crc kubenswrapper[4758]: E1203 18:30:33.747807 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerName="dnsmasq-dns" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.747826 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerName="dnsmasq-dns" Dec 03 18:30:33 crc kubenswrapper[4758]: E1203 18:30:33.747840 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerName="init" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.747846 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerName="init" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.748032 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9aad7c6-1569-4524-9f1e-c8240fd4ad51" containerName="dnsmasq-dns" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.748606 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.772291 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pfhjl"] Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.803575 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bnkk\" (UniqueName: \"kubernetes.io/projected/d1dd1a04-1178-4995-9efa-d843f24e8700-kube-api-access-5bnkk\") pod \"neutron-db-create-pfhjl\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.803645 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1dd1a04-1178-4995-9efa-d843f24e8700-operator-scripts\") pod \"neutron-db-create-pfhjl\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.850265 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-da08-account-create-update-ts4jj"] Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.851430 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.854235 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.863561 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-da08-account-create-update-ts4jj"] Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.904795 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bnkk\" (UniqueName: \"kubernetes.io/projected/d1dd1a04-1178-4995-9efa-d843f24e8700-kube-api-access-5bnkk\") pod \"neutron-db-create-pfhjl\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.904895 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1dd1a04-1178-4995-9efa-d843f24e8700-operator-scripts\") pod \"neutron-db-create-pfhjl\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.905591 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1dd1a04-1178-4995-9efa-d843f24e8700-operator-scripts\") pod \"neutron-db-create-pfhjl\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:33 crc kubenswrapper[4758]: I1203 18:30:33.922713 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bnkk\" (UniqueName: \"kubernetes.io/projected/d1dd1a04-1178-4995-9efa-d843f24e8700-kube-api-access-5bnkk\") pod \"neutron-db-create-pfhjl\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.006410 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35fa890-62ae-4f36-84d4-44cb9dafde45-operator-scripts\") pod \"neutron-da08-account-create-update-ts4jj\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.006559 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnk4w\" (UniqueName: \"kubernetes.io/projected/c35fa890-62ae-4f36-84d4-44cb9dafde45-kube-api-access-cnk4w\") pod \"neutron-da08-account-create-update-ts4jj\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.071046 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.107724 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnk4w\" (UniqueName: \"kubernetes.io/projected/c35fa890-62ae-4f36-84d4-44cb9dafde45-kube-api-access-cnk4w\") pod \"neutron-da08-account-create-update-ts4jj\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.107830 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35fa890-62ae-4f36-84d4-44cb9dafde45-operator-scripts\") pod \"neutron-da08-account-create-update-ts4jj\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.108667 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35fa890-62ae-4f36-84d4-44cb9dafde45-operator-scripts\") pod \"neutron-da08-account-create-update-ts4jj\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.124266 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnk4w\" (UniqueName: \"kubernetes.io/projected/c35fa890-62ae-4f36-84d4-44cb9dafde45-kube-api-access-cnk4w\") pod \"neutron-da08-account-create-update-ts4jj\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.171503 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.507189 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pfhjl"] Dec 03 18:30:34 crc kubenswrapper[4758]: W1203 18:30:34.512403 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1dd1a04_1178_4995_9efa_d843f24e8700.slice/crio-0231843bcff93cd19aece074d594f37ca0c4fd74c02320dd8ddba63de86b9290 WatchSource:0}: Error finding container 0231843bcff93cd19aece074d594f37ca0c4fd74c02320dd8ddba63de86b9290: Status 404 returned error can't find the container with id 0231843bcff93cd19aece074d594f37ca0c4fd74c02320dd8ddba63de86b9290 Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.649556 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-da08-account-create-update-ts4jj"] Dec 03 18:30:34 crc kubenswrapper[4758]: W1203 18:30:34.649960 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc35fa890_62ae_4f36_84d4_44cb9dafde45.slice/crio-0bbb667e2ae1df469be092268149b6fc29dddb6feddfbab9a2f489346456981b WatchSource:0}: Error finding container 0bbb667e2ae1df469be092268149b6fc29dddb6feddfbab9a2f489346456981b: Status 404 returned error can't find the container with id 0bbb667e2ae1df469be092268149b6fc29dddb6feddfbab9a2f489346456981b Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.722770 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da08-account-create-update-ts4jj" event={"ID":"c35fa890-62ae-4f36-84d4-44cb9dafde45","Type":"ContainerStarted","Data":"0bbb667e2ae1df469be092268149b6fc29dddb6feddfbab9a2f489346456981b"} Dec 03 18:30:34 crc kubenswrapper[4758]: I1203 18:30:34.724080 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pfhjl" event={"ID":"d1dd1a04-1178-4995-9efa-d843f24e8700","Type":"ContainerStarted","Data":"0231843bcff93cd19aece074d594f37ca0c4fd74c02320dd8ddba63de86b9290"} Dec 03 18:30:35 crc kubenswrapper[4758]: I1203 18:30:35.732974 4758 generic.go:334] "Generic (PLEG): container finished" podID="c35fa890-62ae-4f36-84d4-44cb9dafde45" containerID="ce44fe8deaa6b02a4616b2d7a016824c7a4b59f50dbebe2495485b924d55501e" exitCode=0 Dec 03 18:30:35 crc kubenswrapper[4758]: I1203 18:30:35.733029 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da08-account-create-update-ts4jj" event={"ID":"c35fa890-62ae-4f36-84d4-44cb9dafde45","Type":"ContainerDied","Data":"ce44fe8deaa6b02a4616b2d7a016824c7a4b59f50dbebe2495485b924d55501e"} Dec 03 18:30:35 crc kubenswrapper[4758]: I1203 18:30:35.734861 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1dd1a04-1178-4995-9efa-d843f24e8700" containerID="4ad5b88232968457a3211d702994039837d6b4ce2209d4b6e178c85feb85cd26" exitCode=0 Dec 03 18:30:35 crc kubenswrapper[4758]: I1203 18:30:35.734899 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pfhjl" event={"ID":"d1dd1a04-1178-4995-9efa-d843f24e8700","Type":"ContainerDied","Data":"4ad5b88232968457a3211d702994039837d6b4ce2209d4b6e178c85feb85cd26"} Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.099456 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.106943 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.156589 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35fa890-62ae-4f36-84d4-44cb9dafde45-operator-scripts\") pod \"c35fa890-62ae-4f36-84d4-44cb9dafde45\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.156639 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1dd1a04-1178-4995-9efa-d843f24e8700-operator-scripts\") pod \"d1dd1a04-1178-4995-9efa-d843f24e8700\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.156725 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnk4w\" (UniqueName: \"kubernetes.io/projected/c35fa890-62ae-4f36-84d4-44cb9dafde45-kube-api-access-cnk4w\") pod \"c35fa890-62ae-4f36-84d4-44cb9dafde45\" (UID: \"c35fa890-62ae-4f36-84d4-44cb9dafde45\") " Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.156763 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bnkk\" (UniqueName: \"kubernetes.io/projected/d1dd1a04-1178-4995-9efa-d843f24e8700-kube-api-access-5bnkk\") pod \"d1dd1a04-1178-4995-9efa-d843f24e8700\" (UID: \"d1dd1a04-1178-4995-9efa-d843f24e8700\") " Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.157447 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1dd1a04-1178-4995-9efa-d843f24e8700-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1dd1a04-1178-4995-9efa-d843f24e8700" (UID: "d1dd1a04-1178-4995-9efa-d843f24e8700"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.157461 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35fa890-62ae-4f36-84d4-44cb9dafde45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c35fa890-62ae-4f36-84d4-44cb9dafde45" (UID: "c35fa890-62ae-4f36-84d4-44cb9dafde45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.163332 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35fa890-62ae-4f36-84d4-44cb9dafde45-kube-api-access-cnk4w" (OuterVolumeSpecName: "kube-api-access-cnk4w") pod "c35fa890-62ae-4f36-84d4-44cb9dafde45" (UID: "c35fa890-62ae-4f36-84d4-44cb9dafde45"). InnerVolumeSpecName "kube-api-access-cnk4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.164998 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1dd1a04-1178-4995-9efa-d843f24e8700-kube-api-access-5bnkk" (OuterVolumeSpecName: "kube-api-access-5bnkk") pod "d1dd1a04-1178-4995-9efa-d843f24e8700" (UID: "d1dd1a04-1178-4995-9efa-d843f24e8700"). InnerVolumeSpecName "kube-api-access-5bnkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.258059 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1dd1a04-1178-4995-9efa-d843f24e8700-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.258105 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnk4w\" (UniqueName: \"kubernetes.io/projected/c35fa890-62ae-4f36-84d4-44cb9dafde45-kube-api-access-cnk4w\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.258120 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bnkk\" (UniqueName: \"kubernetes.io/projected/d1dd1a04-1178-4995-9efa-d843f24e8700-kube-api-access-5bnkk\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.258131 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35fa890-62ae-4f36-84d4-44cb9dafde45-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.751962 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da08-account-create-update-ts4jj" event={"ID":"c35fa890-62ae-4f36-84d4-44cb9dafde45","Type":"ContainerDied","Data":"0bbb667e2ae1df469be092268149b6fc29dddb6feddfbab9a2f489346456981b"} Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.751998 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bbb667e2ae1df469be092268149b6fc29dddb6feddfbab9a2f489346456981b" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.752044 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da08-account-create-update-ts4jj" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.754960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pfhjl" event={"ID":"d1dd1a04-1178-4995-9efa-d843f24e8700","Type":"ContainerDied","Data":"0231843bcff93cd19aece074d594f37ca0c4fd74c02320dd8ddba63de86b9290"} Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.754991 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0231843bcff93cd19aece074d594f37ca0c4fd74c02320dd8ddba63de86b9290" Dec 03 18:30:37 crc kubenswrapper[4758]: I1203 18:30:37.754998 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pfhjl" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.124504 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9gb7v"] Dec 03 18:30:39 crc kubenswrapper[4758]: E1203 18:30:39.125657 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35fa890-62ae-4f36-84d4-44cb9dafde45" containerName="mariadb-account-create-update" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.125673 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35fa890-62ae-4f36-84d4-44cb9dafde45" containerName="mariadb-account-create-update" Dec 03 18:30:39 crc kubenswrapper[4758]: E1203 18:30:39.125731 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dd1a04-1178-4995-9efa-d843f24e8700" containerName="mariadb-database-create" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.125740 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dd1a04-1178-4995-9efa-d843f24e8700" containerName="mariadb-database-create" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.126864 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1dd1a04-1178-4995-9efa-d843f24e8700" containerName="mariadb-database-create" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.126902 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35fa890-62ae-4f36-84d4-44cb9dafde45" containerName="mariadb-account-create-update" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.128399 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.159981 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tz9q7" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.160299 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.164166 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.204867 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9gb7v"] Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.296202 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqbm8\" (UniqueName: \"kubernetes.io/projected/76331360-c8e3-48e2-a71d-1313e699ecd9-kube-api-access-tqbm8\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.296253 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-combined-ca-bundle\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.296325 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-config\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.398504 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqbm8\" (UniqueName: \"kubernetes.io/projected/76331360-c8e3-48e2-a71d-1313e699ecd9-kube-api-access-tqbm8\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.398567 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-combined-ca-bundle\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.398638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-config\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.404124 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-combined-ca-bundle\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.404170 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-config\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.415344 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqbm8\" (UniqueName: \"kubernetes.io/projected/76331360-c8e3-48e2-a71d-1313e699ecd9-kube-api-access-tqbm8\") pod \"neutron-db-sync-9gb7v\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.501908 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:39 crc kubenswrapper[4758]: I1203 18:30:39.938740 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9gb7v"] Dec 03 18:30:40 crc kubenswrapper[4758]: I1203 18:30:40.779627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9gb7v" event={"ID":"76331360-c8e3-48e2-a71d-1313e699ecd9","Type":"ContainerStarted","Data":"e4b7f12e3376095d8a766754ee6f69323f464a37d6ba96f406d01a63d66f8ecc"} Dec 03 18:30:40 crc kubenswrapper[4758]: I1203 18:30:40.780339 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9gb7v" event={"ID":"76331360-c8e3-48e2-a71d-1313e699ecd9","Type":"ContainerStarted","Data":"cff2c1fe5f4d892dff8680c8572b1056717184135fe63f33f6cb7714dc9828ac"} Dec 03 18:30:40 crc kubenswrapper[4758]: I1203 18:30:40.802172 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9gb7v" podStartSLOduration=1.802151359 podStartE2EDuration="1.802151359s" podCreationTimestamp="2025-12-03 18:30:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:40.79510315 +0000 UTC m=+5695.996480011" watchObservedRunningTime="2025-12-03 18:30:40.802151359 +0000 UTC m=+5696.003528220" Dec 03 18:30:45 crc kubenswrapper[4758]: I1203 18:30:45.821414 4758 generic.go:334] "Generic (PLEG): container finished" podID="76331360-c8e3-48e2-a71d-1313e699ecd9" containerID="e4b7f12e3376095d8a766754ee6f69323f464a37d6ba96f406d01a63d66f8ecc" exitCode=0 Dec 03 18:30:45 crc kubenswrapper[4758]: I1203 18:30:45.821487 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9gb7v" event={"ID":"76331360-c8e3-48e2-a71d-1313e699ecd9","Type":"ContainerDied","Data":"e4b7f12e3376095d8a766754ee6f69323f464a37d6ba96f406d01a63d66f8ecc"} Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.161418 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.338963 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-combined-ca-bundle\") pod \"76331360-c8e3-48e2-a71d-1313e699ecd9\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.339511 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-config\") pod \"76331360-c8e3-48e2-a71d-1313e699ecd9\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.339571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqbm8\" (UniqueName: \"kubernetes.io/projected/76331360-c8e3-48e2-a71d-1313e699ecd9-kube-api-access-tqbm8\") pod \"76331360-c8e3-48e2-a71d-1313e699ecd9\" (UID: \"76331360-c8e3-48e2-a71d-1313e699ecd9\") " Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.344314 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76331360-c8e3-48e2-a71d-1313e699ecd9-kube-api-access-tqbm8" (OuterVolumeSpecName: "kube-api-access-tqbm8") pod "76331360-c8e3-48e2-a71d-1313e699ecd9" (UID: "76331360-c8e3-48e2-a71d-1313e699ecd9"). InnerVolumeSpecName "kube-api-access-tqbm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.394339 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76331360-c8e3-48e2-a71d-1313e699ecd9" (UID: "76331360-c8e3-48e2-a71d-1313e699ecd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.400281 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-config" (OuterVolumeSpecName: "config") pod "76331360-c8e3-48e2-a71d-1313e699ecd9" (UID: "76331360-c8e3-48e2-a71d-1313e699ecd9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.441531 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.441573 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqbm8\" (UniqueName: \"kubernetes.io/projected/76331360-c8e3-48e2-a71d-1313e699ecd9-kube-api-access-tqbm8\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.441588 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76331360-c8e3-48e2-a71d-1313e699ecd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.839766 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9gb7v" event={"ID":"76331360-c8e3-48e2-a71d-1313e699ecd9","Type":"ContainerDied","Data":"cff2c1fe5f4d892dff8680c8572b1056717184135fe63f33f6cb7714dc9828ac"} Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.839821 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cff2c1fe5f4d892dff8680c8572b1056717184135fe63f33f6cb7714dc9828ac" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.839825 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9gb7v" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.986577 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c467dc697-n7l99"] Dec 03 18:30:47 crc kubenswrapper[4758]: E1203 18:30:47.987144 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76331360-c8e3-48e2-a71d-1313e699ecd9" containerName="neutron-db-sync" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.987182 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="76331360-c8e3-48e2-a71d-1313e699ecd9" containerName="neutron-db-sync" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.987430 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="76331360-c8e3-48e2-a71d-1313e699ecd9" containerName="neutron-db-sync" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.988695 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:47 crc kubenswrapper[4758]: I1203 18:30:47.994091 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c467dc697-n7l99"] Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.133154 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d7df67465-q66s8"] Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.134739 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.138169 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.138185 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.138466 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tz9q7" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.140596 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d7df67465-q66s8"] Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.152737 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-sb\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.152773 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-dns-svc\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.152834 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-nb\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.152850 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnzkc\" (UniqueName: \"kubernetes.io/projected/bc771b29-799a-4fd0-8e48-32f0cc433b85-kube-api-access-xnzkc\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.152911 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-config\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.253861 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-combined-ca-bundle\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.253947 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-nb\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.253966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnzkc\" (UniqueName: \"kubernetes.io/projected/bc771b29-799a-4fd0-8e48-32f0cc433b85-kube-api-access-xnzkc\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.254012 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-httpd-config\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.254034 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-config\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.254073 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-config\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.254105 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-sb\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.254121 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-dns-svc\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.254137 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6br7\" (UniqueName: \"kubernetes.io/projected/4870a274-a501-43ce-89f6-6d974d05ce8c-kube-api-access-x6br7\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.255559 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-nb\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.255841 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-config\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.256358 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-sb\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.256404 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-dns-svc\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.272540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnzkc\" (UniqueName: \"kubernetes.io/projected/bc771b29-799a-4fd0-8e48-32f0cc433b85-kube-api-access-xnzkc\") pod \"dnsmasq-dns-6c467dc697-n7l99\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.312030 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.356075 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-httpd-config\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.356146 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-config\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.356242 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6br7\" (UniqueName: \"kubernetes.io/projected/4870a274-a501-43ce-89f6-6d974d05ce8c-kube-api-access-x6br7\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.356304 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-combined-ca-bundle\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.359617 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-httpd-config\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.364494 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-config\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.364551 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4870a274-a501-43ce-89f6-6d974d05ce8c-combined-ca-bundle\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.374518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6br7\" (UniqueName: \"kubernetes.io/projected/4870a274-a501-43ce-89f6-6d974d05ce8c-kube-api-access-x6br7\") pod \"neutron-5d7df67465-q66s8\" (UID: \"4870a274-a501-43ce-89f6-6d974d05ce8c\") " pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.452041 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.835781 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c467dc697-n7l99"] Dec 03 18:30:48 crc kubenswrapper[4758]: I1203 18:30:48.848673 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" event={"ID":"bc771b29-799a-4fd0-8e48-32f0cc433b85","Type":"ContainerStarted","Data":"cb0aaf0cfc1aa1e0bf24e2295f7f91c7aba672f8d928103c22b1dce8227ac952"} Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.046021 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d7df67465-q66s8"] Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.859793 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d7df67465-q66s8" event={"ID":"4870a274-a501-43ce-89f6-6d974d05ce8c","Type":"ContainerStarted","Data":"21aabf8dffef97cd7fc5411b05a8b5edef25e6ccf5668e098e4b36b3e1b4ebce"} Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.860765 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d7df67465-q66s8" event={"ID":"4870a274-a501-43ce-89f6-6d974d05ce8c","Type":"ContainerStarted","Data":"a9a7d9042ece140465dbb8919099e1d56f30e0d5a306aed3141226cbc30bb0ab"} Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.860795 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.860850 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d7df67465-q66s8" event={"ID":"4870a274-a501-43ce-89f6-6d974d05ce8c","Type":"ContainerStarted","Data":"b2933a15c3834b2c950fa17844e332b8e038af09b52e7f79b79c8edef5c44b93"} Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.861623 4758 generic.go:334] "Generic (PLEG): container finished" podID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerID="c4557fe9f1c683b3bfb290f44a68ff42fcadadd41374bf6843b2ed7b87de8f5a" exitCode=0 Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.861664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" event={"ID":"bc771b29-799a-4fd0-8e48-32f0cc433b85","Type":"ContainerDied","Data":"c4557fe9f1c683b3bfb290f44a68ff42fcadadd41374bf6843b2ed7b87de8f5a"} Dec 03 18:30:49 crc kubenswrapper[4758]: I1203 18:30:49.904980 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d7df67465-q66s8" podStartSLOduration=1.904955044 podStartE2EDuration="1.904955044s" podCreationTimestamp="2025-12-03 18:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:49.891054272 +0000 UTC m=+5705.092431143" watchObservedRunningTime="2025-12-03 18:30:49.904955044 +0000 UTC m=+5705.106331905" Dec 03 18:30:50 crc kubenswrapper[4758]: I1203 18:30:50.884206 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" event={"ID":"bc771b29-799a-4fd0-8e48-32f0cc433b85","Type":"ContainerStarted","Data":"2c8d0cc9fc88bd1c2364a3dec00e634d67ffb87876839fbf99c771bc63915855"} Dec 03 18:30:50 crc kubenswrapper[4758]: I1203 18:30:50.905989 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:50 crc kubenswrapper[4758]: I1203 18:30:50.945974 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" podStartSLOduration=3.945947834 podStartE2EDuration="3.945947834s" podCreationTimestamp="2025-12-03 18:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:50.933742112 +0000 UTC m=+5706.135118983" watchObservedRunningTime="2025-12-03 18:30:50.945947834 +0000 UTC m=+5706.147324705" Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.313948 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.380954 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb"] Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.381713 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="dnsmasq-dns" containerID="cri-o://2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3" gracePeriod=10 Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.443341 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.33:5353: connect: connection refused" Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.919489 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.948197 4758 generic.go:334] "Generic (PLEG): container finished" podID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerID="2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3" exitCode=0 Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.948240 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" event={"ID":"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1","Type":"ContainerDied","Data":"2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3"} Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.948266 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" event={"ID":"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1","Type":"ContainerDied","Data":"df8bbb851a441d02e52554620c9fc6823ddab38fa66b911bb42372c14abf0840"} Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.948281 4758 scope.go:117] "RemoveContainer" containerID="2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3" Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.948384 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb" Dec 03 18:30:58 crc kubenswrapper[4758]: I1203 18:30:58.970997 4758 scope.go:117] "RemoveContainer" containerID="da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.016294 4758 scope.go:117] "RemoveContainer" containerID="2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3" Dec 03 18:30:59 crc kubenswrapper[4758]: E1203 18:30:59.020266 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3\": container with ID starting with 2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3 not found: ID does not exist" containerID="2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.020304 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3"} err="failed to get container status \"2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3\": rpc error: code = NotFound desc = could not find container \"2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3\": container with ID starting with 2e095f6d11edce7bec97fcd96210cf0a9918ddd1bfe89b4eba82d9c0025bcae3 not found: ID does not exist" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.020327 4758 scope.go:117] "RemoveContainer" containerID="da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33" Dec 03 18:30:59 crc kubenswrapper[4758]: E1203 18:30:59.029817 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33\": container with ID starting with da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33 not found: ID does not exist" containerID="da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.029847 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33"} err="failed to get container status \"da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33\": rpc error: code = NotFound desc = could not find container \"da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33\": container with ID starting with da898509f057d288927feb6a41b1ea837780692f69782164b96ad4a0c96cfb33 not found: ID does not exist" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.062052 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-config\") pod \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.062184 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-sb\") pod \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.062334 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzshv\" (UniqueName: \"kubernetes.io/projected/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-kube-api-access-tzshv\") pod \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.062382 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-nb\") pod \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.062493 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-dns-svc\") pod \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\" (UID: \"54f4b2be-71a6-42eb-9aba-dddc4b3a33a1\") " Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.068301 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-kube-api-access-tzshv" (OuterVolumeSpecName: "kube-api-access-tzshv") pod "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" (UID: "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1"). InnerVolumeSpecName "kube-api-access-tzshv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.142805 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" (UID: "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.146113 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" (UID: "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.148146 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-config" (OuterVolumeSpecName: "config") pod "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" (UID: "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.164743 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" (UID: "54f4b2be-71a6-42eb-9aba-dddc4b3a33a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.164850 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.164883 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.164898 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.164909 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.164923 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzshv\" (UniqueName: \"kubernetes.io/projected/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1-kube-api-access-tzshv\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.279040 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb"] Dec 03 18:30:59 crc kubenswrapper[4758]: I1203 18:30:59.292652 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bf6b7dd8c-sr8fb"] Dec 03 18:31:01 crc kubenswrapper[4758]: I1203 18:31:01.138897 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" path="/var/lib/kubelet/pods/54f4b2be-71a6-42eb-9aba-dddc4b3a33a1/volumes" Dec 03 18:31:18 crc kubenswrapper[4758]: I1203 18:31:18.461772 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d7df67465-q66s8" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.691263 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9hgcn"] Dec 03 18:31:25 crc kubenswrapper[4758]: E1203 18:31:25.692310 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="init" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.692329 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="init" Dec 03 18:31:25 crc kubenswrapper[4758]: E1203 18:31:25.692343 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="dnsmasq-dns" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.692351 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="dnsmasq-dns" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.692554 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="54f4b2be-71a6-42eb-9aba-dddc4b3a33a1" containerName="dnsmasq-dns" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.693335 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.700286 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9hgcn"] Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.791367 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b074-account-create-update-j4m2k"] Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.792870 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.795458 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.800782 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b074-account-create-update-j4m2k"] Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.818987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9563ce-efff-453a-8c17-0704001669ba-operator-scripts\") pod \"glance-db-create-9hgcn\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.819165 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7w7h\" (UniqueName: \"kubernetes.io/projected/8c9563ce-efff-453a-8c17-0704001669ba-kube-api-access-f7w7h\") pod \"glance-db-create-9hgcn\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.921577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-operator-scripts\") pod \"glance-b074-account-create-update-j4m2k\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.921641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9563ce-efff-453a-8c17-0704001669ba-operator-scripts\") pod \"glance-db-create-9hgcn\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.921740 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7w7h\" (UniqueName: \"kubernetes.io/projected/8c9563ce-efff-453a-8c17-0704001669ba-kube-api-access-f7w7h\") pod \"glance-db-create-9hgcn\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.921820 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqfwz\" (UniqueName: \"kubernetes.io/projected/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-kube-api-access-rqfwz\") pod \"glance-b074-account-create-update-j4m2k\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.922706 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9563ce-efff-453a-8c17-0704001669ba-operator-scripts\") pod \"glance-db-create-9hgcn\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:25 crc kubenswrapper[4758]: I1203 18:31:25.942373 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7w7h\" (UniqueName: \"kubernetes.io/projected/8c9563ce-efff-453a-8c17-0704001669ba-kube-api-access-f7w7h\") pod \"glance-db-create-9hgcn\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.011495 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.022994 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-operator-scripts\") pod \"glance-b074-account-create-update-j4m2k\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.023153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqfwz\" (UniqueName: \"kubernetes.io/projected/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-kube-api-access-rqfwz\") pod \"glance-b074-account-create-update-j4m2k\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.023655 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-operator-scripts\") pod \"glance-b074-account-create-update-j4m2k\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.039116 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqfwz\" (UniqueName: \"kubernetes.io/projected/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-kube-api-access-rqfwz\") pod \"glance-b074-account-create-update-j4m2k\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.113667 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.448799 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9hgcn"] Dec 03 18:31:26 crc kubenswrapper[4758]: I1203 18:31:26.601089 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b074-account-create-update-j4m2k"] Dec 03 18:31:26 crc kubenswrapper[4758]: W1203 18:31:26.602429 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dbb51fa_cc6e_4d84_a488_c67712eeb4f3.slice/crio-d6aa2c3566b1b529cad793fdedb9ed721ce64fa9d62e5eb94e5d61a49a5ba8e0 WatchSource:0}: Error finding container d6aa2c3566b1b529cad793fdedb9ed721ce64fa9d62e5eb94e5d61a49a5ba8e0: Status 404 returned error can't find the container with id d6aa2c3566b1b529cad793fdedb9ed721ce64fa9d62e5eb94e5d61a49a5ba8e0 Dec 03 18:31:27 crc kubenswrapper[4758]: I1203 18:31:27.185073 4758 generic.go:334] "Generic (PLEG): container finished" podID="2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" containerID="467e787a7a3dafe0bb1f9481f0a73f819ed5bfdc06681a73212291467ff1628e" exitCode=0 Dec 03 18:31:27 crc kubenswrapper[4758]: I1203 18:31:27.185250 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b074-account-create-update-j4m2k" event={"ID":"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3","Type":"ContainerDied","Data":"467e787a7a3dafe0bb1f9481f0a73f819ed5bfdc06681a73212291467ff1628e"} Dec 03 18:31:27 crc kubenswrapper[4758]: I1203 18:31:27.185370 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b074-account-create-update-j4m2k" event={"ID":"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3","Type":"ContainerStarted","Data":"d6aa2c3566b1b529cad793fdedb9ed721ce64fa9d62e5eb94e5d61a49a5ba8e0"} Dec 03 18:31:27 crc kubenswrapper[4758]: I1203 18:31:27.186775 4758 generic.go:334] "Generic (PLEG): container finished" podID="8c9563ce-efff-453a-8c17-0704001669ba" containerID="b5532523573bc468a8cc42820ae56e455235eb52b6ee507c2ab8308b02ccc779" exitCode=0 Dec 03 18:31:27 crc kubenswrapper[4758]: I1203 18:31:27.186816 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9hgcn" event={"ID":"8c9563ce-efff-453a-8c17-0704001669ba","Type":"ContainerDied","Data":"b5532523573bc468a8cc42820ae56e455235eb52b6ee507c2ab8308b02ccc779"} Dec 03 18:31:27 crc kubenswrapper[4758]: I1203 18:31:27.186841 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9hgcn" event={"ID":"8c9563ce-efff-453a-8c17-0704001669ba","Type":"ContainerStarted","Data":"6803cd461b153ec268f842998d144d14a3ff58473f53a2d4c070614c61939a95"} Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.535399 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.634948 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.676817 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7w7h\" (UniqueName: \"kubernetes.io/projected/8c9563ce-efff-453a-8c17-0704001669ba-kube-api-access-f7w7h\") pod \"8c9563ce-efff-453a-8c17-0704001669ba\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.676970 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9563ce-efff-453a-8c17-0704001669ba-operator-scripts\") pod \"8c9563ce-efff-453a-8c17-0704001669ba\" (UID: \"8c9563ce-efff-453a-8c17-0704001669ba\") " Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.678466 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9563ce-efff-453a-8c17-0704001669ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8c9563ce-efff-453a-8c17-0704001669ba" (UID: "8c9563ce-efff-453a-8c17-0704001669ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.688854 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c9563ce-efff-453a-8c17-0704001669ba-kube-api-access-f7w7h" (OuterVolumeSpecName: "kube-api-access-f7w7h") pod "8c9563ce-efff-453a-8c17-0704001669ba" (UID: "8c9563ce-efff-453a-8c17-0704001669ba"). InnerVolumeSpecName "kube-api-access-f7w7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.778182 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-operator-scripts\") pod \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.778265 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqfwz\" (UniqueName: \"kubernetes.io/projected/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-kube-api-access-rqfwz\") pod \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\" (UID: \"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3\") " Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.778606 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7w7h\" (UniqueName: \"kubernetes.io/projected/8c9563ce-efff-453a-8c17-0704001669ba-kube-api-access-f7w7h\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.778617 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9563ce-efff-453a-8c17-0704001669ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.778782 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" (UID: "2dbb51fa-cc6e-4d84-a488-c67712eeb4f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.780851 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-kube-api-access-rqfwz" (OuterVolumeSpecName: "kube-api-access-rqfwz") pod "2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" (UID: "2dbb51fa-cc6e-4d84-a488-c67712eeb4f3"). InnerVolumeSpecName "kube-api-access-rqfwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.880755 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:28 crc kubenswrapper[4758]: I1203 18:31:28.880884 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqfwz\" (UniqueName: \"kubernetes.io/projected/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3-kube-api-access-rqfwz\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:29 crc kubenswrapper[4758]: I1203 18:31:29.216092 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9hgcn" Dec 03 18:31:29 crc kubenswrapper[4758]: I1203 18:31:29.216122 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9hgcn" event={"ID":"8c9563ce-efff-453a-8c17-0704001669ba","Type":"ContainerDied","Data":"6803cd461b153ec268f842998d144d14a3ff58473f53a2d4c070614c61939a95"} Dec 03 18:31:29 crc kubenswrapper[4758]: I1203 18:31:29.216668 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6803cd461b153ec268f842998d144d14a3ff58473f53a2d4c070614c61939a95" Dec 03 18:31:29 crc kubenswrapper[4758]: I1203 18:31:29.219747 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b074-account-create-update-j4m2k" event={"ID":"2dbb51fa-cc6e-4d84-a488-c67712eeb4f3","Type":"ContainerDied","Data":"d6aa2c3566b1b529cad793fdedb9ed721ce64fa9d62e5eb94e5d61a49a5ba8e0"} Dec 03 18:31:29 crc kubenswrapper[4758]: I1203 18:31:29.219785 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6aa2c3566b1b529cad793fdedb9ed721ce64fa9d62e5eb94e5d61a49a5ba8e0" Dec 03 18:31:29 crc kubenswrapper[4758]: I1203 18:31:29.220873 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b074-account-create-update-j4m2k" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.724024 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-k4w98"] Dec 03 18:31:31 crc kubenswrapper[4758]: E1203 18:31:31.724648 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" containerName="mariadb-account-create-update" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.724660 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" containerName="mariadb-account-create-update" Dec 03 18:31:31 crc kubenswrapper[4758]: E1203 18:31:31.724712 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9563ce-efff-453a-8c17-0704001669ba" containerName="mariadb-database-create" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.724718 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9563ce-efff-453a-8c17-0704001669ba" containerName="mariadb-database-create" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.724874 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9563ce-efff-453a-8c17-0704001669ba" containerName="mariadb-database-create" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.724902 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" containerName="mariadb-account-create-update" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.725473 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.728606 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.728888 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cx5sq" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.742946 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-k4w98"] Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.843449 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f9gf\" (UniqueName: \"kubernetes.io/projected/91dd638a-93de-4683-9e4a-115bb621cd38-kube-api-access-6f9gf\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.843494 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-config-data\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.843513 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-db-sync-config-data\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.843569 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-combined-ca-bundle\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.945472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f9gf\" (UniqueName: \"kubernetes.io/projected/91dd638a-93de-4683-9e4a-115bb621cd38-kube-api-access-6f9gf\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.945515 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-config-data\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.945535 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-db-sync-config-data\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.945589 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-combined-ca-bundle\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.952020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-db-sync-config-data\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.952114 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-config-data\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.958791 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-combined-ca-bundle\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:31 crc kubenswrapper[4758]: I1203 18:31:31.968118 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f9gf\" (UniqueName: \"kubernetes.io/projected/91dd638a-93de-4683-9e4a-115bb621cd38-kube-api-access-6f9gf\") pod \"glance-db-sync-k4w98\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:32 crc kubenswrapper[4758]: I1203 18:31:32.049127 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:32 crc kubenswrapper[4758]: I1203 18:31:32.629214 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-k4w98"] Dec 03 18:31:33 crc kubenswrapper[4758]: I1203 18:31:33.260979 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k4w98" event={"ID":"91dd638a-93de-4683-9e4a-115bb621cd38","Type":"ContainerStarted","Data":"2f114b2c763b993886f79ab64e17c1978c8f3fff8ad14e435968b492daa39e76"} Dec 03 18:31:33 crc kubenswrapper[4758]: I1203 18:31:33.261520 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k4w98" event={"ID":"91dd638a-93de-4683-9e4a-115bb621cd38","Type":"ContainerStarted","Data":"25f2efbfde35ec729a167e997ad4ddacd0f6807169f7d4df15f538970babc940"} Dec 03 18:31:33 crc kubenswrapper[4758]: I1203 18:31:33.284762 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-k4w98" podStartSLOduration=2.284738769 podStartE2EDuration="2.284738769s" podCreationTimestamp="2025-12-03 18:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:33.279728901 +0000 UTC m=+5748.481105782" watchObservedRunningTime="2025-12-03 18:31:33.284738769 +0000 UTC m=+5748.486115630" Dec 03 18:31:37 crc kubenswrapper[4758]: I1203 18:31:37.295495 4758 generic.go:334] "Generic (PLEG): container finished" podID="91dd638a-93de-4683-9e4a-115bb621cd38" containerID="2f114b2c763b993886f79ab64e17c1978c8f3fff8ad14e435968b492daa39e76" exitCode=0 Dec 03 18:31:37 crc kubenswrapper[4758]: I1203 18:31:37.295590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k4w98" event={"ID":"91dd638a-93de-4683-9e4a-115bb621cd38","Type":"ContainerDied","Data":"2f114b2c763b993886f79ab64e17c1978c8f3fff8ad14e435968b492daa39e76"} Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.734116 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.873938 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-combined-ca-bundle\") pod \"91dd638a-93de-4683-9e4a-115bb621cd38\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.873981 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-config-data\") pod \"91dd638a-93de-4683-9e4a-115bb621cd38\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.874070 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-db-sync-config-data\") pod \"91dd638a-93de-4683-9e4a-115bb621cd38\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.874143 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f9gf\" (UniqueName: \"kubernetes.io/projected/91dd638a-93de-4683-9e4a-115bb621cd38-kube-api-access-6f9gf\") pod \"91dd638a-93de-4683-9e4a-115bb621cd38\" (UID: \"91dd638a-93de-4683-9e4a-115bb621cd38\") " Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.880500 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "91dd638a-93de-4683-9e4a-115bb621cd38" (UID: "91dd638a-93de-4683-9e4a-115bb621cd38"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.886876 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91dd638a-93de-4683-9e4a-115bb621cd38-kube-api-access-6f9gf" (OuterVolumeSpecName: "kube-api-access-6f9gf") pod "91dd638a-93de-4683-9e4a-115bb621cd38" (UID: "91dd638a-93de-4683-9e4a-115bb621cd38"). InnerVolumeSpecName "kube-api-access-6f9gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.901975 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91dd638a-93de-4683-9e4a-115bb621cd38" (UID: "91dd638a-93de-4683-9e4a-115bb621cd38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.948868 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-config-data" (OuterVolumeSpecName: "config-data") pod "91dd638a-93de-4683-9e4a-115bb621cd38" (UID: "91dd638a-93de-4683-9e4a-115bb621cd38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.976501 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.976530 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.976541 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/91dd638a-93de-4683-9e4a-115bb621cd38-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:38 crc kubenswrapper[4758]: I1203 18:31:38.976552 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f9gf\" (UniqueName: \"kubernetes.io/projected/91dd638a-93de-4683-9e4a-115bb621cd38-kube-api-access-6f9gf\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.330858 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k4w98" event={"ID":"91dd638a-93de-4683-9e4a-115bb621cd38","Type":"ContainerDied","Data":"25f2efbfde35ec729a167e997ad4ddacd0f6807169f7d4df15f538970babc940"} Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.330935 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25f2efbfde35ec729a167e997ad4ddacd0f6807169f7d4df15f538970babc940" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.331062 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k4w98" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.664587 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64f88c4b89-gtmzq"] Dec 03 18:31:39 crc kubenswrapper[4758]: E1203 18:31:39.664995 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91dd638a-93de-4683-9e4a-115bb621cd38" containerName="glance-db-sync" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.665014 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91dd638a-93de-4683-9e4a-115bb621cd38" containerName="glance-db-sync" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.665170 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="91dd638a-93de-4683-9e4a-115bb621cd38" containerName="glance-db-sync" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.666019 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.680824 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f88c4b89-gtmzq"] Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.690333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-nb\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.690716 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-sb\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.690782 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-config\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.690823 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rfq9\" (UniqueName: \"kubernetes.io/projected/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-kube-api-access-2rfq9\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.690873 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-dns-svc\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.721876 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.723573 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.734196 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.734489 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.734634 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.734878 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cx5sq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.761708 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.793637 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-dns-svc\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.793719 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-nb\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.793790 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-sb\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.793870 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-config\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.793907 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rfq9\" (UniqueName: \"kubernetes.io/projected/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-kube-api-access-2rfq9\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.796196 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-dns-svc\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.797052 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-nb\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.804905 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-config\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.805350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-sb\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.838811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rfq9\" (UniqueName: \"kubernetes.io/projected/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-kube-api-access-2rfq9\") pod \"dnsmasq-dns-64f88c4b89-gtmzq\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.895958 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.895998 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-ceph\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.896034 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-scripts\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.896067 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfzc6\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-kube-api-access-qfzc6\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.896127 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-logs\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.896161 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.896193 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-config-data\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.990645 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997501 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-logs\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997562 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997600 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-config-data\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997735 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997769 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-ceph\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997799 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-scripts\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.997835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfzc6\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-kube-api-access-qfzc6\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.998694 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-logs\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:39 crc kubenswrapper[4758]: I1203 18:31:39.998965 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.003195 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-config-data\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.009469 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.014280 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-ceph\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.015032 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-scripts\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.024471 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfzc6\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-kube-api-access-qfzc6\") pod \"glance-default-external-api-0\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.044484 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.046546 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.058185 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.058296 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.070462 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.201413 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.201660 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.201718 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.201762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.201959 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.201988 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx4hp\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-kube-api-access-tx4hp\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.202128 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.303953 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.304042 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.304080 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.304109 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.304140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx4hp\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-kube-api-access-tx4hp\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.304178 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.304332 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.306208 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.307020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.311047 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.311285 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.311722 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.312270 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.323878 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx4hp\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-kube-api-access-tx4hp\") pod \"glance-default-internal-api-0\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.505551 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.519373 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f88c4b89-gtmzq"] Dec 03 18:31:40 crc kubenswrapper[4758]: I1203 18:31:40.759165 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:40 crc kubenswrapper[4758]: W1203 18:31:40.761529 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5b3aef0_83dc_4743_a984_cce22ad69cf8.slice/crio-ed38ed3b8d4a560f13560ac8b6da9b39a830a23048984868317a34b77fd6c259 WatchSource:0}: Error finding container ed38ed3b8d4a560f13560ac8b6da9b39a830a23048984868317a34b77fd6c259: Status 404 returned error can't find the container with id ed38ed3b8d4a560f13560ac8b6da9b39a830a23048984868317a34b77fd6c259 Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.029239 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.084549 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:41 crc kubenswrapper[4758]: W1203 18:31:41.093288 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode70442ba_fdb6_449d_b5a2_a6a7bb93f2e3.slice/crio-a6133c49b566c9aa0958eba9c4ad6a10359de0e3fc99eadacfd2afd81fc2ada2 WatchSource:0}: Error finding container a6133c49b566c9aa0958eba9c4ad6a10359de0e3fc99eadacfd2afd81fc2ada2: Status 404 returned error can't find the container with id a6133c49b566c9aa0958eba9c4ad6a10359de0e3fc99eadacfd2afd81fc2ada2 Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.370227 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5b3aef0-83dc-4743-a984-cce22ad69cf8","Type":"ContainerStarted","Data":"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7"} Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.370279 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5b3aef0-83dc-4743-a984-cce22ad69cf8","Type":"ContainerStarted","Data":"ed38ed3b8d4a560f13560ac8b6da9b39a830a23048984868317a34b77fd6c259"} Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.386883 4758 generic.go:334] "Generic (PLEG): container finished" podID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerID="bcd1706f8793d6ce6669ea773816c9f8ee8ccd69e16e80561b7e6b8b97c4bba6" exitCode=0 Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.386963 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" event={"ID":"4dbae1d2-d2c3-432f-8907-d49e37ad6d05","Type":"ContainerDied","Data":"bcd1706f8793d6ce6669ea773816c9f8ee8ccd69e16e80561b7e6b8b97c4bba6"} Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.386993 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" event={"ID":"4dbae1d2-d2c3-432f-8907-d49e37ad6d05","Type":"ContainerStarted","Data":"fa7fac40dbf1c9afa8b915c6c3211816e5baeb905ea7eb83266159a497316b81"} Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.391439 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3","Type":"ContainerStarted","Data":"a6133c49b566c9aa0958eba9c4ad6a10359de0e3fc99eadacfd2afd81fc2ada2"} Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.394333 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:31:41 crc kubenswrapper[4758]: I1203 18:31:41.394390 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.401101 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" event={"ID":"4dbae1d2-d2c3-432f-8907-d49e37ad6d05","Type":"ContainerStarted","Data":"d98ff59a09225f515d3bf96df59cd9b21d67de9ded9c513efbb186af1b8a0de7"} Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.403118 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.404907 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3","Type":"ContainerStarted","Data":"33bbccb03351afe8817a78d772523bcabb4696d5bd00a6df1b7ea9ffbe01d500"} Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.408094 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5b3aef0-83dc-4743-a984-cce22ad69cf8","Type":"ContainerStarted","Data":"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2"} Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.408228 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-log" containerID="cri-o://d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7" gracePeriod=30 Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.408399 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-httpd" containerID="cri-o://4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2" gracePeriod=30 Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.434702 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" podStartSLOduration=3.434665592 podStartE2EDuration="3.434665592s" podCreationTimestamp="2025-12-03 18:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:42.427005535 +0000 UTC m=+5757.628382436" watchObservedRunningTime="2025-12-03 18:31:42.434665592 +0000 UTC m=+5757.636042463" Dec 03 18:31:42 crc kubenswrapper[4758]: I1203 18:31:42.492314 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.492292729 podStartE2EDuration="3.492292729s" podCreationTimestamp="2025-12-03 18:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:42.489204238 +0000 UTC m=+5757.690581099" watchObservedRunningTime="2025-12-03 18:31:42.492292729 +0000 UTC m=+5757.693669590" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.024836 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.105958 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172266 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-logs\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172327 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-httpd-run\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172357 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-config-data\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172380 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfzc6\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-kube-api-access-qfzc6\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172406 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-scripts\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172419 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-combined-ca-bundle\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172476 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-ceph\") pod \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\" (UID: \"b5b3aef0-83dc-4743-a984-cce22ad69cf8\") " Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172814 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-logs" (OuterVolumeSpecName: "logs") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.172943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.180107 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.180139 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5b3aef0-83dc-4743-a984-cce22ad69cf8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.192967 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-ceph" (OuterVolumeSpecName: "ceph") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.206886 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-kube-api-access-qfzc6" (OuterVolumeSpecName: "kube-api-access-qfzc6") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "kube-api-access-qfzc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.215013 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-scripts" (OuterVolumeSpecName: "scripts") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.245001 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.269865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-config-data" (OuterVolumeSpecName: "config-data") pod "b5b3aef0-83dc-4743-a984-cce22ad69cf8" (UID: "b5b3aef0-83dc-4743-a984-cce22ad69cf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.281300 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.281336 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfzc6\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-kube-api-access-qfzc6\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.281347 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.281357 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b3aef0-83dc-4743-a984-cce22ad69cf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.281368 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5b3aef0-83dc-4743-a984-cce22ad69cf8-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.421177 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3","Type":"ContainerStarted","Data":"27e7180f6ea2b35347a7c38bcf0bac39c7fcdffc66c39f430e3d15ad67e7dfd4"} Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.422941 4758 generic.go:334] "Generic (PLEG): container finished" podID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerID="4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2" exitCode=0 Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.422967 4758 generic.go:334] "Generic (PLEG): container finished" podID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerID="d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7" exitCode=143 Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.423690 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5b3aef0-83dc-4743-a984-cce22ad69cf8","Type":"ContainerDied","Data":"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2"} Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.423748 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5b3aef0-83dc-4743-a984-cce22ad69cf8","Type":"ContainerDied","Data":"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7"} Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.423651 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.423779 4758 scope.go:117] "RemoveContainer" containerID="4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.423760 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5b3aef0-83dc-4743-a984-cce22ad69cf8","Type":"ContainerDied","Data":"ed38ed3b8d4a560f13560ac8b6da9b39a830a23048984868317a34b77fd6c259"} Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.453418 4758 scope.go:117] "RemoveContainer" containerID="d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.464312 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.464295445 podStartE2EDuration="4.464295445s" podCreationTimestamp="2025-12-03 18:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:43.45200636 +0000 UTC m=+5758.653383221" watchObservedRunningTime="2025-12-03 18:31:43.464295445 +0000 UTC m=+5758.665672306" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.472636 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.473299 4758 scope.go:117] "RemoveContainer" containerID="4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2" Dec 03 18:31:43 crc kubenswrapper[4758]: E1203 18:31:43.473811 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2\": container with ID starting with 4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2 not found: ID does not exist" containerID="4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.473847 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2"} err="failed to get container status \"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2\": rpc error: code = NotFound desc = could not find container \"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2\": container with ID starting with 4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2 not found: ID does not exist" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.473873 4758 scope.go:117] "RemoveContainer" containerID="d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7" Dec 03 18:31:43 crc kubenswrapper[4758]: E1203 18:31:43.474187 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7\": container with ID starting with d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7 not found: ID does not exist" containerID="d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.474226 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7"} err="failed to get container status \"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7\": rpc error: code = NotFound desc = could not find container \"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7\": container with ID starting with d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7 not found: ID does not exist" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.474244 4758 scope.go:117] "RemoveContainer" containerID="4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.478442 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2"} err="failed to get container status \"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2\": rpc error: code = NotFound desc = could not find container \"4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2\": container with ID starting with 4f24d23fc694680e684118157a62d052d57d55713b732f612a0aaf326bcb29f2 not found: ID does not exist" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.478495 4758 scope.go:117] "RemoveContainer" containerID="d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.478943 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7"} err="failed to get container status \"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7\": rpc error: code = NotFound desc = could not find container \"d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7\": container with ID starting with d8bd15b8041cdf12bdd1cf7512d39458a2bfec7b2b87552a2949e62c418109e7 not found: ID does not exist" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.483428 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.495458 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:43 crc kubenswrapper[4758]: E1203 18:31:43.495839 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-log" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.495853 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-log" Dec 03 18:31:43 crc kubenswrapper[4758]: E1203 18:31:43.495876 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-httpd" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.495882 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-httpd" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.496050 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-httpd" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.496066 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" containerName="glance-log" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.496919 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.502405 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.533612 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586708 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586812 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-config-data\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586853 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xrd4\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-kube-api-access-2xrd4\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586896 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-scripts\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586938 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-logs\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586965 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-ceph\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.586999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689174 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689266 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-config-data\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xrd4\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-kube-api-access-2xrd4\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689348 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-scripts\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689389 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-logs\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-ceph\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.689455 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.690298 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-logs\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.690364 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.693394 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-scripts\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.698562 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.698598 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-ceph\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.699204 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-config-data\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.708510 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xrd4\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-kube-api-access-2xrd4\") pod \"glance-default-external-api-0\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " pod="openstack/glance-default-external-api-0" Dec 03 18:31:43 crc kubenswrapper[4758]: I1203 18:31:43.856276 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:31:44 crc kubenswrapper[4758]: I1203 18:31:44.385877 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:31:44 crc kubenswrapper[4758]: I1203 18:31:44.433967 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f386b68a-b4be-459f-a332-f93ec438ddcb","Type":"ContainerStarted","Data":"8231711a22fe55f0bc62b377117f1b3579ac340571e3c177c27febc049134a3a"} Dec 03 18:31:44 crc kubenswrapper[4758]: I1203 18:31:44.437969 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-log" containerID="cri-o://33bbccb03351afe8817a78d772523bcabb4696d5bd00a6df1b7ea9ffbe01d500" gracePeriod=30 Dec 03 18:31:44 crc kubenswrapper[4758]: I1203 18:31:44.438069 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-httpd" containerID="cri-o://27e7180f6ea2b35347a7c38bcf0bac39c7fcdffc66c39f430e3d15ad67e7dfd4" gracePeriod=30 Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.130595 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b3aef0-83dc-4743-a984-cce22ad69cf8" path="/var/lib/kubelet/pods/b5b3aef0-83dc-4743-a984-cce22ad69cf8/volumes" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.456652 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f386b68a-b4be-459f-a332-f93ec438ddcb","Type":"ContainerStarted","Data":"303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38"} Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.459713 4758 generic.go:334] "Generic (PLEG): container finished" podID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerID="27e7180f6ea2b35347a7c38bcf0bac39c7fcdffc66c39f430e3d15ad67e7dfd4" exitCode=0 Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.459743 4758 generic.go:334] "Generic (PLEG): container finished" podID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerID="33bbccb03351afe8817a78d772523bcabb4696d5bd00a6df1b7ea9ffbe01d500" exitCode=143 Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.459766 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3","Type":"ContainerDied","Data":"27e7180f6ea2b35347a7c38bcf0bac39c7fcdffc66c39f430e3d15ad67e7dfd4"} Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.459792 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3","Type":"ContainerDied","Data":"33bbccb03351afe8817a78d772523bcabb4696d5bd00a6df1b7ea9ffbe01d500"} Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.613906 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639297 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-combined-ca-bundle\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639355 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-config-data\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639488 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-scripts\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639560 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx4hp\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-kube-api-access-tx4hp\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-ceph\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639638 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-httpd-run\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.639655 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-logs\") pod \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\" (UID: \"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3\") " Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.640276 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-logs" (OuterVolumeSpecName: "logs") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.642853 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.681926 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-ceph" (OuterVolumeSpecName: "ceph") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.684019 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-kube-api-access-tx4hp" (OuterVolumeSpecName: "kube-api-access-tx4hp") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "kube-api-access-tx4hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.687006 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-scripts" (OuterVolumeSpecName: "scripts") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.688176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.721827 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-config-data" (OuterVolumeSpecName: "config-data") pod "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" (UID: "e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741416 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741451 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx4hp\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-kube-api-access-tx4hp\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741464 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741472 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741481 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741489 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:45 crc kubenswrapper[4758]: I1203 18:31:45.741497 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.471941 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.471936 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3","Type":"ContainerDied","Data":"a6133c49b566c9aa0958eba9c4ad6a10359de0e3fc99eadacfd2afd81fc2ada2"} Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.473725 4758 scope.go:117] "RemoveContainer" containerID="27e7180f6ea2b35347a7c38bcf0bac39c7fcdffc66c39f430e3d15ad67e7dfd4" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.473942 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f386b68a-b4be-459f-a332-f93ec438ddcb","Type":"ContainerStarted","Data":"0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179"} Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.506482 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.5064614069999998 podStartE2EDuration="3.506461407s" podCreationTimestamp="2025-12-03 18:31:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:46.503106388 +0000 UTC m=+5761.704483289" watchObservedRunningTime="2025-12-03 18:31:46.506461407 +0000 UTC m=+5761.707838268" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.508606 4758 scope.go:117] "RemoveContainer" containerID="33bbccb03351afe8817a78d772523bcabb4696d5bd00a6df1b7ea9ffbe01d500" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.524176 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.541136 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.550096 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:46 crc kubenswrapper[4758]: E1203 18:31:46.550426 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-log" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.550443 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-log" Dec 03 18:31:46 crc kubenswrapper[4758]: E1203 18:31:46.550470 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-httpd" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.550476 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-httpd" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.550630 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-log" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.550651 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" containerName="glance-httpd" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.551577 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553128 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-logs\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553313 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx2cj\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-kube-api-access-hx2cj\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553409 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553526 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553605 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.553896 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.555294 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.562729 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.656013 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.656361 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.656383 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-logs\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.656409 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx2cj\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-kube-api-access-hx2cj\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.658336 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-logs\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.658412 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.658440 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.659150 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.659051 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.660215 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.661075 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.661721 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.662184 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.678102 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx2cj\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-kube-api-access-hx2cj\") pod \"glance-default-internal-api-0\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:31:46 crc kubenswrapper[4758]: I1203 18:31:46.885296 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:47 crc kubenswrapper[4758]: I1203 18:31:47.129007 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3" path="/var/lib/kubelet/pods/e70442ba-fdb6-449d-b5a2-a6a7bb93f2e3/volumes" Dec 03 18:31:47 crc kubenswrapper[4758]: I1203 18:31:47.429023 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:47 crc kubenswrapper[4758]: W1203 18:31:47.436912 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode611edbe_5bdd_4061_baaf_fc92217e4eaf.slice/crio-ff7736751c3a239f6d246c17c997ba3ddde2283ad1194218bc8b0237c4da8b29 WatchSource:0}: Error finding container ff7736751c3a239f6d246c17c997ba3ddde2283ad1194218bc8b0237c4da8b29: Status 404 returned error can't find the container with id ff7736751c3a239f6d246c17c997ba3ddde2283ad1194218bc8b0237c4da8b29 Dec 03 18:31:47 crc kubenswrapper[4758]: I1203 18:31:47.496117 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e611edbe-5bdd-4061-baaf-fc92217e4eaf","Type":"ContainerStarted","Data":"ff7736751c3a239f6d246c17c997ba3ddde2283ad1194218bc8b0237c4da8b29"} Dec 03 18:31:48 crc kubenswrapper[4758]: I1203 18:31:48.508086 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e611edbe-5bdd-4061-baaf-fc92217e4eaf","Type":"ContainerStarted","Data":"cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783"} Dec 03 18:31:48 crc kubenswrapper[4758]: I1203 18:31:48.508630 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e611edbe-5bdd-4061-baaf-fc92217e4eaf","Type":"ContainerStarted","Data":"17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98"} Dec 03 18:31:48 crc kubenswrapper[4758]: I1203 18:31:48.536808 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.536782294 podStartE2EDuration="2.536782294s" podCreationTimestamp="2025-12-03 18:31:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:48.529078456 +0000 UTC m=+5763.730455327" watchObservedRunningTime="2025-12-03 18:31:48.536782294 +0000 UTC m=+5763.738159165" Dec 03 18:31:49 crc kubenswrapper[4758]: I1203 18:31:49.993455 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.064556 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c467dc697-n7l99"] Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.064946 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerName="dnsmasq-dns" containerID="cri-o://2c8d0cc9fc88bd1c2364a3dec00e634d67ffb87876839fbf99c771bc63915855" gracePeriod=10 Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.524042 4758 generic.go:334] "Generic (PLEG): container finished" podID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerID="2c8d0cc9fc88bd1c2364a3dec00e634d67ffb87876839fbf99c771bc63915855" exitCode=0 Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.524346 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" event={"ID":"bc771b29-799a-4fd0-8e48-32f0cc433b85","Type":"ContainerDied","Data":"2c8d0cc9fc88bd1c2364a3dec00e634d67ffb87876839fbf99c771bc63915855"} Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.524424 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" event={"ID":"bc771b29-799a-4fd0-8e48-32f0cc433b85","Type":"ContainerDied","Data":"cb0aaf0cfc1aa1e0bf24e2295f7f91c7aba672f8d928103c22b1dce8227ac952"} Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.524436 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb0aaf0cfc1aa1e0bf24e2295f7f91c7aba672f8d928103c22b1dce8227ac952" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.585370 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.721620 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnzkc\" (UniqueName: \"kubernetes.io/projected/bc771b29-799a-4fd0-8e48-32f0cc433b85-kube-api-access-xnzkc\") pod \"bc771b29-799a-4fd0-8e48-32f0cc433b85\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.721750 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-config\") pod \"bc771b29-799a-4fd0-8e48-32f0cc433b85\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.721798 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-dns-svc\") pod \"bc771b29-799a-4fd0-8e48-32f0cc433b85\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.721836 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-sb\") pod \"bc771b29-799a-4fd0-8e48-32f0cc433b85\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.721950 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-nb\") pod \"bc771b29-799a-4fd0-8e48-32f0cc433b85\" (UID: \"bc771b29-799a-4fd0-8e48-32f0cc433b85\") " Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.731944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc771b29-799a-4fd0-8e48-32f0cc433b85-kube-api-access-xnzkc" (OuterVolumeSpecName: "kube-api-access-xnzkc") pod "bc771b29-799a-4fd0-8e48-32f0cc433b85" (UID: "bc771b29-799a-4fd0-8e48-32f0cc433b85"). InnerVolumeSpecName "kube-api-access-xnzkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.767809 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc771b29-799a-4fd0-8e48-32f0cc433b85" (UID: "bc771b29-799a-4fd0-8e48-32f0cc433b85"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.771362 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc771b29-799a-4fd0-8e48-32f0cc433b85" (UID: "bc771b29-799a-4fd0-8e48-32f0cc433b85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.775268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-config" (OuterVolumeSpecName: "config") pod "bc771b29-799a-4fd0-8e48-32f0cc433b85" (UID: "bc771b29-799a-4fd0-8e48-32f0cc433b85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.775454 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc771b29-799a-4fd0-8e48-32f0cc433b85" (UID: "bc771b29-799a-4fd0-8e48-32f0cc433b85"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.823467 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnzkc\" (UniqueName: \"kubernetes.io/projected/bc771b29-799a-4fd0-8e48-32f0cc433b85-kube-api-access-xnzkc\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.823496 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.823508 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.823517 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:50 crc kubenswrapper[4758]: I1203 18:31:50.823527 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc771b29-799a-4fd0-8e48-32f0cc433b85-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4758]: I1203 18:31:51.535002 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c467dc697-n7l99" Dec 03 18:31:51 crc kubenswrapper[4758]: I1203 18:31:51.564464 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c467dc697-n7l99"] Dec 03 18:31:51 crc kubenswrapper[4758]: I1203 18:31:51.571723 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c467dc697-n7l99"] Dec 03 18:31:53 crc kubenswrapper[4758]: I1203 18:31:53.133327 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" path="/var/lib/kubelet/pods/bc771b29-799a-4fd0-8e48-32f0cc433b85/volumes" Dec 03 18:31:53 crc kubenswrapper[4758]: I1203 18:31:53.856629 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 18:31:53 crc kubenswrapper[4758]: I1203 18:31:53.856750 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 18:31:53 crc kubenswrapper[4758]: I1203 18:31:53.896751 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 18:31:53 crc kubenswrapper[4758]: I1203 18:31:53.910279 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 18:31:54 crc kubenswrapper[4758]: I1203 18:31:54.574497 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 18:31:54 crc kubenswrapper[4758]: I1203 18:31:54.574600 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 18:31:56 crc kubenswrapper[4758]: I1203 18:31:56.885568 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:56 crc kubenswrapper[4758]: I1203 18:31:56.885881 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:56 crc kubenswrapper[4758]: I1203 18:31:56.938596 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:56 crc kubenswrapper[4758]: I1203 18:31:56.938916 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:57 crc kubenswrapper[4758]: I1203 18:31:57.544981 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 18:31:57 crc kubenswrapper[4758]: I1203 18:31:57.545103 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 18:31:57 crc kubenswrapper[4758]: I1203 18:31:57.550458 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 18:31:57 crc kubenswrapper[4758]: I1203 18:31:57.610833 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:57 crc kubenswrapper[4758]: I1203 18:31:57.610884 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:59 crc kubenswrapper[4758]: I1203 18:31:59.618976 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:59 crc kubenswrapper[4758]: I1203 18:31:59.621005 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.445700 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-shwqh"] Dec 03 18:32:07 crc kubenswrapper[4758]: E1203 18:32:07.446559 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerName="init" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.446572 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerName="init" Dec 03 18:32:07 crc kubenswrapper[4758]: E1203 18:32:07.446587 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerName="dnsmasq-dns" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.446593 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerName="dnsmasq-dns" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.446773 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc771b29-799a-4fd0-8e48-32f0cc433b85" containerName="dnsmasq-dns" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.447550 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.458161 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-shwqh"] Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.486765 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52fdf77e-873b-47f1-9c59-6f46226c43a6-operator-scripts\") pod \"placement-db-create-shwqh\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.486910 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s24rq\" (UniqueName: \"kubernetes.io/projected/52fdf77e-873b-47f1-9c59-6f46226c43a6-kube-api-access-s24rq\") pod \"placement-db-create-shwqh\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.547757 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bd30-account-create-update-r7bsr"] Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.548773 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.557255 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.575653 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd30-account-create-update-r7bsr"] Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.588272 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s24rq\" (UniqueName: \"kubernetes.io/projected/52fdf77e-873b-47f1-9c59-6f46226c43a6-kube-api-access-s24rq\") pod \"placement-db-create-shwqh\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.588334 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-operator-scripts\") pod \"placement-bd30-account-create-update-r7bsr\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.588371 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52fdf77e-873b-47f1-9c59-6f46226c43a6-operator-scripts\") pod \"placement-db-create-shwqh\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.588392 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq8x5\" (UniqueName: \"kubernetes.io/projected/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-kube-api-access-dq8x5\") pod \"placement-bd30-account-create-update-r7bsr\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.589155 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52fdf77e-873b-47f1-9c59-6f46226c43a6-operator-scripts\") pod \"placement-db-create-shwqh\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.605778 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s24rq\" (UniqueName: \"kubernetes.io/projected/52fdf77e-873b-47f1-9c59-6f46226c43a6-kube-api-access-s24rq\") pod \"placement-db-create-shwqh\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.689601 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-operator-scripts\") pod \"placement-bd30-account-create-update-r7bsr\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.689934 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq8x5\" (UniqueName: \"kubernetes.io/projected/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-kube-api-access-dq8x5\") pod \"placement-bd30-account-create-update-r7bsr\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.690547 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-operator-scripts\") pod \"placement-bd30-account-create-update-r7bsr\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.706367 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq8x5\" (UniqueName: \"kubernetes.io/projected/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-kube-api-access-dq8x5\") pod \"placement-bd30-account-create-update-r7bsr\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.810088 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-shwqh" Dec 03 18:32:07 crc kubenswrapper[4758]: I1203 18:32:07.868884 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.309823 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-shwqh"] Dec 03 18:32:08 crc kubenswrapper[4758]: W1203 18:32:08.310816 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52fdf77e_873b_47f1_9c59_6f46226c43a6.slice/crio-f38487d7d40c47bd916c3e438321b67aa28f5094da7e801dbae7152f83812af1 WatchSource:0}: Error finding container f38487d7d40c47bd916c3e438321b67aa28f5094da7e801dbae7152f83812af1: Status 404 returned error can't find the container with id f38487d7d40c47bd916c3e438321b67aa28f5094da7e801dbae7152f83812af1 Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.357035 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd30-account-create-update-r7bsr"] Dec 03 18:32:08 crc kubenswrapper[4758]: W1203 18:32:08.358487 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab563bad_7d6e_4c2e_bccd_b882230dcc9c.slice/crio-d0000b64f89e4adab3eeca09d4f4247786b0d40f6c44b5e894a16e5948b64c76 WatchSource:0}: Error finding container d0000b64f89e4adab3eeca09d4f4247786b0d40f6c44b5e894a16e5948b64c76: Status 404 returned error can't find the container with id d0000b64f89e4adab3eeca09d4f4247786b0d40f6c44b5e894a16e5948b64c76 Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.709455 4758 generic.go:334] "Generic (PLEG): container finished" podID="52fdf77e-873b-47f1-9c59-6f46226c43a6" containerID="5bd75ec17d038ee42069ba402b6f27348155bd2da28cb6a2cc402dd65fdb3df6" exitCode=0 Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.709566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-shwqh" event={"ID":"52fdf77e-873b-47f1-9c59-6f46226c43a6","Type":"ContainerDied","Data":"5bd75ec17d038ee42069ba402b6f27348155bd2da28cb6a2cc402dd65fdb3df6"} Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.710105 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-shwqh" event={"ID":"52fdf77e-873b-47f1-9c59-6f46226c43a6","Type":"ContainerStarted","Data":"f38487d7d40c47bd916c3e438321b67aa28f5094da7e801dbae7152f83812af1"} Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.711598 4758 generic.go:334] "Generic (PLEG): container finished" podID="ab563bad-7d6e-4c2e-bccd-b882230dcc9c" containerID="c3220aded8dc6ef27ba410012855f7bc774d86bb755f01b192172bf67754e6d8" exitCode=0 Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.711667 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd30-account-create-update-r7bsr" event={"ID":"ab563bad-7d6e-4c2e-bccd-b882230dcc9c","Type":"ContainerDied","Data":"c3220aded8dc6ef27ba410012855f7bc774d86bb755f01b192172bf67754e6d8"} Dec 03 18:32:08 crc kubenswrapper[4758]: I1203 18:32:08.711706 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd30-account-create-update-r7bsr" event={"ID":"ab563bad-7d6e-4c2e-bccd-b882230dcc9c","Type":"ContainerStarted","Data":"d0000b64f89e4adab3eeca09d4f4247786b0d40f6c44b5e894a16e5948b64c76"} Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.146349 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.151592 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-shwqh" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.340885 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52fdf77e-873b-47f1-9c59-6f46226c43a6-operator-scripts\") pod \"52fdf77e-873b-47f1-9c59-6f46226c43a6\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.340970 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq8x5\" (UniqueName: \"kubernetes.io/projected/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-kube-api-access-dq8x5\") pod \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.341015 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-operator-scripts\") pod \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\" (UID: \"ab563bad-7d6e-4c2e-bccd-b882230dcc9c\") " Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.341107 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s24rq\" (UniqueName: \"kubernetes.io/projected/52fdf77e-873b-47f1-9c59-6f46226c43a6-kube-api-access-s24rq\") pod \"52fdf77e-873b-47f1-9c59-6f46226c43a6\" (UID: \"52fdf77e-873b-47f1-9c59-6f46226c43a6\") " Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.341705 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab563bad-7d6e-4c2e-bccd-b882230dcc9c" (UID: "ab563bad-7d6e-4c2e-bccd-b882230dcc9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.341967 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52fdf77e-873b-47f1-9c59-6f46226c43a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52fdf77e-873b-47f1-9c59-6f46226c43a6" (UID: "52fdf77e-873b-47f1-9c59-6f46226c43a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.346052 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52fdf77e-873b-47f1-9c59-6f46226c43a6-kube-api-access-s24rq" (OuterVolumeSpecName: "kube-api-access-s24rq") pod "52fdf77e-873b-47f1-9c59-6f46226c43a6" (UID: "52fdf77e-873b-47f1-9c59-6f46226c43a6"). InnerVolumeSpecName "kube-api-access-s24rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.346411 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-kube-api-access-dq8x5" (OuterVolumeSpecName: "kube-api-access-dq8x5") pod "ab563bad-7d6e-4c2e-bccd-b882230dcc9c" (UID: "ab563bad-7d6e-4c2e-bccd-b882230dcc9c"). InnerVolumeSpecName "kube-api-access-dq8x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.443442 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s24rq\" (UniqueName: \"kubernetes.io/projected/52fdf77e-873b-47f1-9c59-6f46226c43a6-kube-api-access-s24rq\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.443481 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52fdf77e-873b-47f1-9c59-6f46226c43a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.443490 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq8x5\" (UniqueName: \"kubernetes.io/projected/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-kube-api-access-dq8x5\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.443499 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab563bad-7d6e-4c2e-bccd-b882230dcc9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.730145 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-shwqh" event={"ID":"52fdf77e-873b-47f1-9c59-6f46226c43a6","Type":"ContainerDied","Data":"f38487d7d40c47bd916c3e438321b67aa28f5094da7e801dbae7152f83812af1"} Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.730444 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f38487d7d40c47bd916c3e438321b67aa28f5094da7e801dbae7152f83812af1" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.730204 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-shwqh" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.732222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd30-account-create-update-r7bsr" event={"ID":"ab563bad-7d6e-4c2e-bccd-b882230dcc9c","Type":"ContainerDied","Data":"d0000b64f89e4adab3eeca09d4f4247786b0d40f6c44b5e894a16e5948b64c76"} Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.732258 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0000b64f89e4adab3eeca09d4f4247786b0d40f6c44b5e894a16e5948b64c76" Dec 03 18:32:10 crc kubenswrapper[4758]: I1203 18:32:10.732340 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd30-account-create-update-r7bsr" Dec 03 18:32:11 crc kubenswrapper[4758]: I1203 18:32:11.394480 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:32:11 crc kubenswrapper[4758]: I1203 18:32:11.394541 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:32:11 crc kubenswrapper[4758]: I1203 18:32:11.423413 4758 scope.go:117] "RemoveContainer" containerID="175644fa55952194731916bbad3b39a444c476323d83d2b88c2f7cd7ee4c525a" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.871897 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5584c87ddf-q6bzc"] Dec 03 18:32:12 crc kubenswrapper[4758]: E1203 18:32:12.872903 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab563bad-7d6e-4c2e-bccd-b882230dcc9c" containerName="mariadb-account-create-update" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.872920 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab563bad-7d6e-4c2e-bccd-b882230dcc9c" containerName="mariadb-account-create-update" Dec 03 18:32:12 crc kubenswrapper[4758]: E1203 18:32:12.872956 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52fdf77e-873b-47f1-9c59-6f46226c43a6" containerName="mariadb-database-create" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.872964 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="52fdf77e-873b-47f1-9c59-6f46226c43a6" containerName="mariadb-database-create" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.873157 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab563bad-7d6e-4c2e-bccd-b882230dcc9c" containerName="mariadb-account-create-update" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.873180 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="52fdf77e-873b-47f1-9c59-6f46226c43a6" containerName="mariadb-database-create" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.874357 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.889357 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5584c87ddf-q6bzc"] Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.920080 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dkgcn"] Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.922852 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.925601 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.926294 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.926354 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-m8dqs" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.937628 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dkgcn"] Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.983988 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-dns-svc\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.984034 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk6p4\" (UniqueName: \"kubernetes.io/projected/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-kube-api-access-mk6p4\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.984094 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-nb\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.984274 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-config\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:12 crc kubenswrapper[4758]: I1203 18:32:12.984346 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-sb\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086467 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-config\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086536 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-config-data\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086565 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-sb\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086602 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6khlz\" (UniqueName: \"kubernetes.io/projected/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-kube-api-access-6khlz\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086692 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-combined-ca-bundle\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086742 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-dns-svc\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086777 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk6p4\" (UniqueName: \"kubernetes.io/projected/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-kube-api-access-mk6p4\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-nb\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-scripts\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.086916 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-logs\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.087812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-dns-svc\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.087849 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-nb\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.087875 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-sb\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.088398 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-config\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.107155 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk6p4\" (UniqueName: \"kubernetes.io/projected/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-kube-api-access-mk6p4\") pod \"dnsmasq-dns-5584c87ddf-q6bzc\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.189053 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-combined-ca-bundle\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.189319 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-scripts\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.189358 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-logs\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.189403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-config-data\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.189447 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6khlz\" (UniqueName: \"kubernetes.io/projected/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-kube-api-access-6khlz\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.191081 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-logs\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.192981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-scripts\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.193568 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.193645 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-combined-ca-bundle\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.193932 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-config-data\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.205929 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6khlz\" (UniqueName: \"kubernetes.io/projected/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-kube-api-access-6khlz\") pod \"placement-db-sync-dkgcn\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.244863 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.673960 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5584c87ddf-q6bzc"] Dec 03 18:32:13 crc kubenswrapper[4758]: W1203 18:32:13.731722 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e0d1b1d_8465_47f6_89ad_a05266e6acaf.slice/crio-8d4baff1767d4982fa2735b672bb43c9a45a51e287e78b57c06fc263c6fc828a WatchSource:0}: Error finding container 8d4baff1767d4982fa2735b672bb43c9a45a51e287e78b57c06fc263c6fc828a: Status 404 returned error can't find the container with id 8d4baff1767d4982fa2735b672bb43c9a45a51e287e78b57c06fc263c6fc828a Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.731912 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dkgcn"] Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.754442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" event={"ID":"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2","Type":"ContainerStarted","Data":"e66b5f4cafbed5224144232d6c2892b23162b94ad643fc03e275b83d14a529d7"} Dec 03 18:32:13 crc kubenswrapper[4758]: I1203 18:32:13.755475 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dkgcn" event={"ID":"6e0d1b1d-8465-47f6-89ad-a05266e6acaf","Type":"ContainerStarted","Data":"8d4baff1767d4982fa2735b672bb43c9a45a51e287e78b57c06fc263c6fc828a"} Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.253313 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-28898"] Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.255845 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.291780 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28898"] Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.418816 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-utilities\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.419164 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-catalog-content\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.419186 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hzjl\" (UniqueName: \"kubernetes.io/projected/6cbfdd97-96fb-4da4-84b0-04d02e209104-kube-api-access-8hzjl\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.520936 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-catalog-content\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.520990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hzjl\" (UniqueName: \"kubernetes.io/projected/6cbfdd97-96fb-4da4-84b0-04d02e209104-kube-api-access-8hzjl\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.521096 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-utilities\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.521519 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-utilities\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.521518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-catalog-content\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.540925 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hzjl\" (UniqueName: \"kubernetes.io/projected/6cbfdd97-96fb-4da4-84b0-04d02e209104-kube-api-access-8hzjl\") pod \"redhat-marketplace-28898\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.596986 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.764286 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dkgcn" event={"ID":"6e0d1b1d-8465-47f6-89ad-a05266e6acaf","Type":"ContainerStarted","Data":"93a6fb1412c9cf87fcc7b623b4ab03be77937a1653811378ce7fdad5ab6c01f4"} Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.767287 4758 generic.go:334] "Generic (PLEG): container finished" podID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerID="9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9" exitCode=0 Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.767390 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" event={"ID":"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2","Type":"ContainerDied","Data":"9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9"} Dec 03 18:32:14 crc kubenswrapper[4758]: I1203 18:32:14.793198 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dkgcn" podStartSLOduration=2.793174289 podStartE2EDuration="2.793174289s" podCreationTimestamp="2025-12-03 18:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:32:14.784023353 +0000 UTC m=+5789.985400204" watchObservedRunningTime="2025-12-03 18:32:14.793174289 +0000 UTC m=+5789.994551150" Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.145600 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28898"] Dec 03 18:32:15 crc kubenswrapper[4758]: W1203 18:32:15.159917 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cbfdd97_96fb_4da4_84b0_04d02e209104.slice/crio-ebc693f4b5ef4bd4ac29297cb3f7dfee23ca6c13649280d09f9572f274506a6d WatchSource:0}: Error finding container ebc693f4b5ef4bd4ac29297cb3f7dfee23ca6c13649280d09f9572f274506a6d: Status 404 returned error can't find the container with id ebc693f4b5ef4bd4ac29297cb3f7dfee23ca6c13649280d09f9572f274506a6d Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.776631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" event={"ID":"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2","Type":"ContainerStarted","Data":"92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717"} Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.777007 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.778556 4758 generic.go:334] "Generic (PLEG): container finished" podID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerID="7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b" exitCode=0 Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.778612 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28898" event={"ID":"6cbfdd97-96fb-4da4-84b0-04d02e209104","Type":"ContainerDied","Data":"7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b"} Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.778636 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28898" event={"ID":"6cbfdd97-96fb-4da4-84b0-04d02e209104","Type":"ContainerStarted","Data":"ebc693f4b5ef4bd4ac29297cb3f7dfee23ca6c13649280d09f9572f274506a6d"} Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.781021 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.782994 4758 generic.go:334] "Generic (PLEG): container finished" podID="6e0d1b1d-8465-47f6-89ad-a05266e6acaf" containerID="93a6fb1412c9cf87fcc7b623b4ab03be77937a1653811378ce7fdad5ab6c01f4" exitCode=0 Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.783051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dkgcn" event={"ID":"6e0d1b1d-8465-47f6-89ad-a05266e6acaf","Type":"ContainerDied","Data":"93a6fb1412c9cf87fcc7b623b4ab03be77937a1653811378ce7fdad5ab6c01f4"} Dec 03 18:32:15 crc kubenswrapper[4758]: I1203 18:32:15.823799 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" podStartSLOduration=3.823776601 podStartE2EDuration="3.823776601s" podCreationTimestamp="2025-12-03 18:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:32:15.80364704 +0000 UTC m=+5791.005023891" watchObservedRunningTime="2025-12-03 18:32:15.823776601 +0000 UTC m=+5791.025153462" Dec 03 18:32:16 crc kubenswrapper[4758]: I1203 18:32:16.795621 4758 generic.go:334] "Generic (PLEG): container finished" podID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerID="aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1" exitCode=0 Dec 03 18:32:16 crc kubenswrapper[4758]: I1203 18:32:16.795728 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28898" event={"ID":"6cbfdd97-96fb-4da4-84b0-04d02e209104","Type":"ContainerDied","Data":"aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1"} Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.194000 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.196318 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-config-data\") pod \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.196367 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6khlz\" (UniqueName: \"kubernetes.io/projected/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-kube-api-access-6khlz\") pod \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.196505 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-combined-ca-bundle\") pod \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.196531 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-scripts\") pod \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.196555 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-logs\") pod \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\" (UID: \"6e0d1b1d-8465-47f6-89ad-a05266e6acaf\") " Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.197041 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-logs" (OuterVolumeSpecName: "logs") pod "6e0d1b1d-8465-47f6-89ad-a05266e6acaf" (UID: "6e0d1b1d-8465-47f6-89ad-a05266e6acaf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.201535 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-scripts" (OuterVolumeSpecName: "scripts") pod "6e0d1b1d-8465-47f6-89ad-a05266e6acaf" (UID: "6e0d1b1d-8465-47f6-89ad-a05266e6acaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.203068 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-kube-api-access-6khlz" (OuterVolumeSpecName: "kube-api-access-6khlz") pod "6e0d1b1d-8465-47f6-89ad-a05266e6acaf" (UID: "6e0d1b1d-8465-47f6-89ad-a05266e6acaf"). InnerVolumeSpecName "kube-api-access-6khlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.203222 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6khlz\" (UniqueName: \"kubernetes.io/projected/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-kube-api-access-6khlz\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.203237 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.203245 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.226394 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-config-data" (OuterVolumeSpecName: "config-data") pod "6e0d1b1d-8465-47f6-89ad-a05266e6acaf" (UID: "6e0d1b1d-8465-47f6-89ad-a05266e6acaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.231050 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e0d1b1d-8465-47f6-89ad-a05266e6acaf" (UID: "6e0d1b1d-8465-47f6-89ad-a05266e6acaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.305147 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.305193 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0d1b1d-8465-47f6-89ad-a05266e6acaf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.815234 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dkgcn" event={"ID":"6e0d1b1d-8465-47f6-89ad-a05266e6acaf","Type":"ContainerDied","Data":"8d4baff1767d4982fa2735b672bb43c9a45a51e287e78b57c06fc263c6fc828a"} Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.815280 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d4baff1767d4982fa2735b672bb43c9a45a51e287e78b57c06fc263c6fc828a" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.815400 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dkgcn" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.988597 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-645b59b84b-kb6kd"] Dec 03 18:32:17 crc kubenswrapper[4758]: E1203 18:32:17.989205 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0d1b1d-8465-47f6-89ad-a05266e6acaf" containerName="placement-db-sync" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.989221 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0d1b1d-8465-47f6-89ad-a05266e6acaf" containerName="placement-db-sync" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.989410 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0d1b1d-8465-47f6-89ad-a05266e6acaf" containerName="placement-db-sync" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.991984 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.997097 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.997169 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-m8dqs" Dec 03 18:32:17 crc kubenswrapper[4758]: I1203 18:32:17.997455 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.006359 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-645b59b84b-kb6kd"] Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.016019 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-config-data\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.016069 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h69pv\" (UniqueName: \"kubernetes.io/projected/99571942-c27b-4b35-8d59-3ba4450b2f3a-kube-api-access-h69pv\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.016100 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99571942-c27b-4b35-8d59-3ba4450b2f3a-logs\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.016132 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-combined-ca-bundle\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.016168 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-scripts\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.116707 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-config-data\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.117035 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h69pv\" (UniqueName: \"kubernetes.io/projected/99571942-c27b-4b35-8d59-3ba4450b2f3a-kube-api-access-h69pv\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.117083 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99571942-c27b-4b35-8d59-3ba4450b2f3a-logs\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.117125 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-combined-ca-bundle\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.117166 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-scripts\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.118241 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99571942-c27b-4b35-8d59-3ba4450b2f3a-logs\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.123718 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-scripts\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.123796 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-combined-ca-bundle\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.123850 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99571942-c27b-4b35-8d59-3ba4450b2f3a-config-data\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.133920 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h69pv\" (UniqueName: \"kubernetes.io/projected/99571942-c27b-4b35-8d59-3ba4450b2f3a-kube-api-access-h69pv\") pod \"placement-645b59b84b-kb6kd\" (UID: \"99571942-c27b-4b35-8d59-3ba4450b2f3a\") " pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.312237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.818301 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-645b59b84b-kb6kd"] Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.828202 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28898" event={"ID":"6cbfdd97-96fb-4da4-84b0-04d02e209104","Type":"ContainerStarted","Data":"5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde"} Dec 03 18:32:18 crc kubenswrapper[4758]: I1203 18:32:18.857959 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-28898" podStartSLOduration=2.93351404 podStartE2EDuration="4.857944319s" podCreationTimestamp="2025-12-03 18:32:14 +0000 UTC" firstStartedPulling="2025-12-03 18:32:15.780708313 +0000 UTC m=+5790.982085194" lastFinishedPulling="2025-12-03 18:32:17.705138562 +0000 UTC m=+5792.906515473" observedRunningTime="2025-12-03 18:32:18.854738293 +0000 UTC m=+5794.056115154" watchObservedRunningTime="2025-12-03 18:32:18.857944319 +0000 UTC m=+5794.059321180" Dec 03 18:32:19 crc kubenswrapper[4758]: I1203 18:32:19.841978 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-645b59b84b-kb6kd" event={"ID":"99571942-c27b-4b35-8d59-3ba4450b2f3a","Type":"ContainerStarted","Data":"0118241f0ded50c43ec1a22104309f96e59331ace9875bbbbd12457a6b6ffcd8"} Dec 03 18:32:19 crc kubenswrapper[4758]: I1203 18:32:19.842815 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:19 crc kubenswrapper[4758]: I1203 18:32:19.842850 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-645b59b84b-kb6kd" event={"ID":"99571942-c27b-4b35-8d59-3ba4450b2f3a","Type":"ContainerStarted","Data":"c53b03d4187cdd1590694d6264e039fe3f41005aaa831bb3adefa94eb95b406e"} Dec 03 18:32:19 crc kubenswrapper[4758]: I1203 18:32:19.842875 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-645b59b84b-kb6kd" event={"ID":"99571942-c27b-4b35-8d59-3ba4450b2f3a","Type":"ContainerStarted","Data":"fa375461b4ff094a67c4dc5d7f079c6852fad0f6270ad13aed8ca8f0e90a5a2f"} Dec 03 18:32:19 crc kubenswrapper[4758]: I1203 18:32:19.842907 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:19 crc kubenswrapper[4758]: I1203 18:32:19.878345 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-645b59b84b-kb6kd" podStartSLOduration=2.878318936 podStartE2EDuration="2.878318936s" podCreationTimestamp="2025-12-03 18:32:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:32:19.866812767 +0000 UTC m=+5795.068189658" watchObservedRunningTime="2025-12-03 18:32:19.878318936 +0000 UTC m=+5795.079695807" Dec 03 18:32:23 crc kubenswrapper[4758]: I1203 18:32:23.194876 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:32:23 crc kubenswrapper[4758]: I1203 18:32:23.258224 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f88c4b89-gtmzq"] Dec 03 18:32:23 crc kubenswrapper[4758]: I1203 18:32:23.258489 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerName="dnsmasq-dns" containerID="cri-o://d98ff59a09225f515d3bf96df59cd9b21d67de9ded9c513efbb186af1b8a0de7" gracePeriod=10 Dec 03 18:32:23 crc kubenswrapper[4758]: I1203 18:32:23.884321 4758 generic.go:334] "Generic (PLEG): container finished" podID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerID="d98ff59a09225f515d3bf96df59cd9b21d67de9ded9c513efbb186af1b8a0de7" exitCode=0 Dec 03 18:32:23 crc kubenswrapper[4758]: I1203 18:32:23.884364 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" event={"ID":"4dbae1d2-d2c3-432f-8907-d49e37ad6d05","Type":"ContainerDied","Data":"d98ff59a09225f515d3bf96df59cd9b21d67de9ded9c513efbb186af1b8a0de7"} Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.192234 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.233337 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-config\") pod \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.233391 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-dns-svc\") pod \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.233466 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-sb\") pod \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.233516 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-nb\") pod \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.234296 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rfq9\" (UniqueName: \"kubernetes.io/projected/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-kube-api-access-2rfq9\") pod \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\" (UID: \"4dbae1d2-d2c3-432f-8907-d49e37ad6d05\") " Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.254865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-kube-api-access-2rfq9" (OuterVolumeSpecName: "kube-api-access-2rfq9") pod "4dbae1d2-d2c3-432f-8907-d49e37ad6d05" (UID: "4dbae1d2-d2c3-432f-8907-d49e37ad6d05"). InnerVolumeSpecName "kube-api-access-2rfq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.275849 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4dbae1d2-d2c3-432f-8907-d49e37ad6d05" (UID: "4dbae1d2-d2c3-432f-8907-d49e37ad6d05"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.277745 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4dbae1d2-d2c3-432f-8907-d49e37ad6d05" (UID: "4dbae1d2-d2c3-432f-8907-d49e37ad6d05"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.284848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4dbae1d2-d2c3-432f-8907-d49e37ad6d05" (UID: "4dbae1d2-d2c3-432f-8907-d49e37ad6d05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.288592 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-config" (OuterVolumeSpecName: "config") pod "4dbae1d2-d2c3-432f-8907-d49e37ad6d05" (UID: "4dbae1d2-d2c3-432f-8907-d49e37ad6d05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.335696 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.335765 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.335785 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.335804 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rfq9\" (UniqueName: \"kubernetes.io/projected/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-kube-api-access-2rfq9\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.335824 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dbae1d2-d2c3-432f-8907-d49e37ad6d05-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.597660 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.597815 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.653244 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.893074 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" event={"ID":"4dbae1d2-d2c3-432f-8907-d49e37ad6d05","Type":"ContainerDied","Data":"fa7fac40dbf1c9afa8b915c6c3211816e5baeb905ea7eb83266159a497316b81"} Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.893105 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f88c4b89-gtmzq" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.893124 4758 scope.go:117] "RemoveContainer" containerID="d98ff59a09225f515d3bf96df59cd9b21d67de9ded9c513efbb186af1b8a0de7" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.926396 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f88c4b89-gtmzq"] Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.926420 4758 scope.go:117] "RemoveContainer" containerID="bcd1706f8793d6ce6669ea773816c9f8ee8ccd69e16e80561b7e6b8b97c4bba6" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.942579 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64f88c4b89-gtmzq"] Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.945800 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:24 crc kubenswrapper[4758]: I1203 18:32:24.993884 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-28898"] Dec 03 18:32:25 crc kubenswrapper[4758]: I1203 18:32:25.127793 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" path="/var/lib/kubelet/pods/4dbae1d2-d2c3-432f-8907-d49e37ad6d05/volumes" Dec 03 18:32:26 crc kubenswrapper[4758]: I1203 18:32:26.913842 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-28898" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="registry-server" containerID="cri-o://5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde" gracePeriod=2 Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.398207 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.490384 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hzjl\" (UniqueName: \"kubernetes.io/projected/6cbfdd97-96fb-4da4-84b0-04d02e209104-kube-api-access-8hzjl\") pod \"6cbfdd97-96fb-4da4-84b0-04d02e209104\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.490699 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-utilities\") pod \"6cbfdd97-96fb-4da4-84b0-04d02e209104\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.490891 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-catalog-content\") pod \"6cbfdd97-96fb-4da4-84b0-04d02e209104\" (UID: \"6cbfdd97-96fb-4da4-84b0-04d02e209104\") " Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.491382 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-utilities" (OuterVolumeSpecName: "utilities") pod "6cbfdd97-96fb-4da4-84b0-04d02e209104" (UID: "6cbfdd97-96fb-4da4-84b0-04d02e209104"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.491923 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.499365 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cbfdd97-96fb-4da4-84b0-04d02e209104-kube-api-access-8hzjl" (OuterVolumeSpecName: "kube-api-access-8hzjl") pod "6cbfdd97-96fb-4da4-84b0-04d02e209104" (UID: "6cbfdd97-96fb-4da4-84b0-04d02e209104"). InnerVolumeSpecName "kube-api-access-8hzjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.513475 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6cbfdd97-96fb-4da4-84b0-04d02e209104" (UID: "6cbfdd97-96fb-4da4-84b0-04d02e209104"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.593340 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbfdd97-96fb-4da4-84b0-04d02e209104-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.593390 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hzjl\" (UniqueName: \"kubernetes.io/projected/6cbfdd97-96fb-4da4-84b0-04d02e209104-kube-api-access-8hzjl\") on node \"crc\" DevicePath \"\"" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.934030 4758 generic.go:334] "Generic (PLEG): container finished" podID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerID="5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde" exitCode=0 Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.934123 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28898" event={"ID":"6cbfdd97-96fb-4da4-84b0-04d02e209104","Type":"ContainerDied","Data":"5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde"} Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.934149 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28898" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.934200 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28898" event={"ID":"6cbfdd97-96fb-4da4-84b0-04d02e209104","Type":"ContainerDied","Data":"ebc693f4b5ef4bd4ac29297cb3f7dfee23ca6c13649280d09f9572f274506a6d"} Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.934239 4758 scope.go:117] "RemoveContainer" containerID="5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde" Dec 03 18:32:27 crc kubenswrapper[4758]: I1203 18:32:27.968038 4758 scope.go:117] "RemoveContainer" containerID="aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.000742 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-28898"] Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.014009 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-28898"] Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.034128 4758 scope.go:117] "RemoveContainer" containerID="7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.053798 4758 scope.go:117] "RemoveContainer" containerID="5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde" Dec 03 18:32:28 crc kubenswrapper[4758]: E1203 18:32:28.054385 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde\": container with ID starting with 5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde not found: ID does not exist" containerID="5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.054422 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde"} err="failed to get container status \"5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde\": rpc error: code = NotFound desc = could not find container \"5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde\": container with ID starting with 5e3b6de9413a28f76ba089aa2a2efbfb63bfdb05d8c13844787c40c4caea4dde not found: ID does not exist" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.054468 4758 scope.go:117] "RemoveContainer" containerID="aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1" Dec 03 18:32:28 crc kubenswrapper[4758]: E1203 18:32:28.054915 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1\": container with ID starting with aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1 not found: ID does not exist" containerID="aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.054955 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1"} err="failed to get container status \"aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1\": rpc error: code = NotFound desc = could not find container \"aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1\": container with ID starting with aa4fb1360970ee94af9e82cfda3c7d15a98378154187eef1901b3aae48df3ff1 not found: ID does not exist" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.054981 4758 scope.go:117] "RemoveContainer" containerID="7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b" Dec 03 18:32:28 crc kubenswrapper[4758]: E1203 18:32:28.055355 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b\": container with ID starting with 7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b not found: ID does not exist" containerID="7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b" Dec 03 18:32:28 crc kubenswrapper[4758]: I1203 18:32:28.055405 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b"} err="failed to get container status \"7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b\": rpc error: code = NotFound desc = could not find container \"7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b\": container with ID starting with 7c93a5229ebbef07c8fc465b793a59a2c80453c7bbe27c51d936be1ccf0f8a1b not found: ID does not exist" Dec 03 18:32:29 crc kubenswrapper[4758]: I1203 18:32:29.125750 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" path="/var/lib/kubelet/pods/6cbfdd97-96fb-4da4-84b0-04d02e209104/volumes" Dec 03 18:32:41 crc kubenswrapper[4758]: I1203 18:32:41.394352 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:32:41 crc kubenswrapper[4758]: I1203 18:32:41.394953 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:32:41 crc kubenswrapper[4758]: I1203 18:32:41.395066 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:32:41 crc kubenswrapper[4758]: I1203 18:32:41.395979 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:32:41 crc kubenswrapper[4758]: I1203 18:32:41.396051 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" gracePeriod=600 Dec 03 18:32:41 crc kubenswrapper[4758]: E1203 18:32:41.516691 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:32:42 crc kubenswrapper[4758]: I1203 18:32:42.134300 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" exitCode=0 Dec 03 18:32:42 crc kubenswrapper[4758]: I1203 18:32:42.134384 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3"} Dec 03 18:32:42 crc kubenswrapper[4758]: I1203 18:32:42.134588 4758 scope.go:117] "RemoveContainer" containerID="ee02f3819ed4276671e01c5782910e87da9be6dd0555ff89c3f0a62d67c490da" Dec 03 18:32:42 crc kubenswrapper[4758]: I1203 18:32:42.135048 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:32:42 crc kubenswrapper[4758]: E1203 18:32:42.135363 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:32:49 crc kubenswrapper[4758]: I1203 18:32:49.496808 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:50 crc kubenswrapper[4758]: I1203 18:32:50.491322 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-645b59b84b-kb6kd" Dec 03 18:32:55 crc kubenswrapper[4758]: I1203 18:32:55.125014 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:32:55 crc kubenswrapper[4758]: E1203 18:32:55.125995 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.071499 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zmnnm"] Dec 03 18:33:05 crc kubenswrapper[4758]: E1203 18:33:05.072567 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="extract-content" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072580 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="extract-content" Dec 03 18:33:05 crc kubenswrapper[4758]: E1203 18:33:05.072591 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerName="dnsmasq-dns" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072597 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerName="dnsmasq-dns" Dec 03 18:33:05 crc kubenswrapper[4758]: E1203 18:33:05.072613 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="extract-utilities" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072620 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="extract-utilities" Dec 03 18:33:05 crc kubenswrapper[4758]: E1203 18:33:05.072637 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerName="init" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072642 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerName="init" Dec 03 18:33:05 crc kubenswrapper[4758]: E1203 18:33:05.072659 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="registry-server" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072691 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="registry-server" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072880 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbfdd97-96fb-4da4-84b0-04d02e209104" containerName="registry-server" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.072904 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dbae1d2-d2c3-432f-8907-d49e37ad6d05" containerName="dnsmasq-dns" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.074208 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.082572 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zmnnm"] Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.185485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-utilities\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.185577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-catalog-content\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.185604 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf7h8\" (UniqueName: \"kubernetes.io/projected/f8c3f795-e3d2-452e-a85e-b38746e7abee-kube-api-access-sf7h8\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.287058 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-catalog-content\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.287105 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf7h8\" (UniqueName: \"kubernetes.io/projected/f8c3f795-e3d2-452e-a85e-b38746e7abee-kube-api-access-sf7h8\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.287200 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-utilities\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.287754 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-catalog-content\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.287787 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-utilities\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.316394 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf7h8\" (UniqueName: \"kubernetes.io/projected/f8c3f795-e3d2-452e-a85e-b38746e7abee-kube-api-access-sf7h8\") pod \"certified-operators-zmnnm\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.435760 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:05 crc kubenswrapper[4758]: I1203 18:33:05.936988 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zmnnm"] Dec 03 18:33:06 crc kubenswrapper[4758]: I1203 18:33:06.337881 4758 generic.go:334] "Generic (PLEG): container finished" podID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerID="b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258" exitCode=0 Dec 03 18:33:06 crc kubenswrapper[4758]: I1203 18:33:06.337926 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmnnm" event={"ID":"f8c3f795-e3d2-452e-a85e-b38746e7abee","Type":"ContainerDied","Data":"b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258"} Dec 03 18:33:06 crc kubenswrapper[4758]: I1203 18:33:06.338228 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmnnm" event={"ID":"f8c3f795-e3d2-452e-a85e-b38746e7abee","Type":"ContainerStarted","Data":"1ed06f7fc8af39a3370d89835e094289c226208d402ebfd9d1ca4a6e86fda7c7"} Dec 03 18:33:07 crc kubenswrapper[4758]: I1203 18:33:07.114831 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:33:07 crc kubenswrapper[4758]: E1203 18:33:07.115247 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.098411 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-kct8l"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.100575 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.112125 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kct8l"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.188741 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-knwsq"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.189772 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.193481 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67htp\" (UniqueName: \"kubernetes.io/projected/3597b615-57a3-47b9-a718-35e266252642-kube-api-access-67htp\") pod \"nova-api-db-create-kct8l\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.193762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3597b615-57a3-47b9-a718-35e266252642-operator-scripts\") pod \"nova-api-db-create-kct8l\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.210135 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-knwsq"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.294919 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-9mv8v"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.295548 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e84029c-4acb-4f87-baee-86fff838a10d-operator-scripts\") pod \"nova-cell0-db-create-knwsq\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.295599 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67htp\" (UniqueName: \"kubernetes.io/projected/3597b615-57a3-47b9-a718-35e266252642-kube-api-access-67htp\") pod \"nova-api-db-create-kct8l\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.295945 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42jzj\" (UniqueName: \"kubernetes.io/projected/4e84029c-4acb-4f87-baee-86fff838a10d-kube-api-access-42jzj\") pod \"nova-cell0-db-create-knwsq\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.296048 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3597b615-57a3-47b9-a718-35e266252642-operator-scripts\") pod \"nova-api-db-create-kct8l\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.296069 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.296635 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3597b615-57a3-47b9-a718-35e266252642-operator-scripts\") pod \"nova-api-db-create-kct8l\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.304618 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-bea5-account-create-update-d5xv2"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.305695 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.309659 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.317178 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9mv8v"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.319510 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67htp\" (UniqueName: \"kubernetes.io/projected/3597b615-57a3-47b9-a718-35e266252642-kube-api-access-67htp\") pod \"nova-api-db-create-kct8l\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.352170 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bea5-account-create-update-d5xv2"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.399102 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwmmc\" (UniqueName: \"kubernetes.io/projected/c2ff9e71-648d-48e2-b428-ee5676abc858-kube-api-access-dwmmc\") pod \"nova-cell1-db-create-9mv8v\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.399337 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e84029c-4acb-4f87-baee-86fff838a10d-operator-scripts\") pod \"nova-cell0-db-create-knwsq\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.399420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42jzj\" (UniqueName: \"kubernetes.io/projected/4e84029c-4acb-4f87-baee-86fff838a10d-kube-api-access-42jzj\") pod \"nova-cell0-db-create-knwsq\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.399502 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27d6cf1-c67d-4b5d-9a54-d1681c949841-operator-scripts\") pod \"nova-api-bea5-account-create-update-d5xv2\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.399583 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwkpz\" (UniqueName: \"kubernetes.io/projected/f27d6cf1-c67d-4b5d-9a54-d1681c949841-kube-api-access-dwkpz\") pod \"nova-api-bea5-account-create-update-d5xv2\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.399712 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ff9e71-648d-48e2-b428-ee5676abc858-operator-scripts\") pod \"nova-cell1-db-create-9mv8v\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.400536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e84029c-4acb-4f87-baee-86fff838a10d-operator-scripts\") pod \"nova-cell0-db-create-knwsq\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.420330 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42jzj\" (UniqueName: \"kubernetes.io/projected/4e84029c-4acb-4f87-baee-86fff838a10d-kube-api-access-42jzj\") pod \"nova-cell0-db-create-knwsq\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.422250 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.496287 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cdcf-account-create-update-bftp6"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.500650 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.501229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ff9e71-648d-48e2-b428-ee5676abc858-operator-scripts\") pod \"nova-cell1-db-create-9mv8v\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.501326 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwmmc\" (UniqueName: \"kubernetes.io/projected/c2ff9e71-648d-48e2-b428-ee5676abc858-kube-api-access-dwmmc\") pod \"nova-cell1-db-create-9mv8v\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.501390 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27d6cf1-c67d-4b5d-9a54-d1681c949841-operator-scripts\") pod \"nova-api-bea5-account-create-update-d5xv2\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.501426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwkpz\" (UniqueName: \"kubernetes.io/projected/f27d6cf1-c67d-4b5d-9a54-d1681c949841-kube-api-access-dwkpz\") pod \"nova-api-bea5-account-create-update-d5xv2\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.502205 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ff9e71-648d-48e2-b428-ee5676abc858-operator-scripts\") pod \"nova-cell1-db-create-9mv8v\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.502243 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27d6cf1-c67d-4b5d-9a54-d1681c949841-operator-scripts\") pod \"nova-api-bea5-account-create-update-d5xv2\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.503714 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.517215 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cdcf-account-create-update-bftp6"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.527406 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwmmc\" (UniqueName: \"kubernetes.io/projected/c2ff9e71-648d-48e2-b428-ee5676abc858-kube-api-access-dwmmc\") pod \"nova-cell1-db-create-9mv8v\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.533291 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwkpz\" (UniqueName: \"kubernetes.io/projected/f27d6cf1-c67d-4b5d-9a54-d1681c949841-kube-api-access-dwkpz\") pod \"nova-api-bea5-account-create-update-d5xv2\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.537611 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.603238 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353f3f28-3228-4a1b-9485-8b5f38247264-operator-scripts\") pod \"nova-cell0-cdcf-account-create-update-bftp6\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.603301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76g85\" (UniqueName: \"kubernetes.io/projected/353f3f28-3228-4a1b-9485-8b5f38247264-kube-api-access-76g85\") pod \"nova-cell0-cdcf-account-create-update-bftp6\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.613486 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.662017 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.701725 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4cde-account-create-update-9bwcq"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.703801 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.704743 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353f3f28-3228-4a1b-9485-8b5f38247264-operator-scripts\") pod \"nova-cell0-cdcf-account-create-update-bftp6\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.704858 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76g85\" (UniqueName: \"kubernetes.io/projected/353f3f28-3228-4a1b-9485-8b5f38247264-kube-api-access-76g85\") pod \"nova-cell0-cdcf-account-create-update-bftp6\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.705594 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353f3f28-3228-4a1b-9485-8b5f38247264-operator-scripts\") pod \"nova-cell0-cdcf-account-create-update-bftp6\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.708049 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.714474 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4cde-account-create-update-9bwcq"] Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.730578 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76g85\" (UniqueName: \"kubernetes.io/projected/353f3f28-3228-4a1b-9485-8b5f38247264-kube-api-access-76g85\") pod \"nova-cell0-cdcf-account-create-update-bftp6\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.806920 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c270177-5d19-4a1a-a965-f2b106c43e26-operator-scripts\") pod \"nova-cell1-4cde-account-create-update-9bwcq\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.807428 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xbr7\" (UniqueName: \"kubernetes.io/projected/5c270177-5d19-4a1a-a965-f2b106c43e26-kube-api-access-7xbr7\") pod \"nova-cell1-4cde-account-create-update-9bwcq\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.819336 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.909390 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xbr7\" (UniqueName: \"kubernetes.io/projected/5c270177-5d19-4a1a-a965-f2b106c43e26-kube-api-access-7xbr7\") pod \"nova-cell1-4cde-account-create-update-9bwcq\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.909943 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c270177-5d19-4a1a-a965-f2b106c43e26-operator-scripts\") pod \"nova-cell1-4cde-account-create-update-9bwcq\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.910793 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c270177-5d19-4a1a-a965-f2b106c43e26-operator-scripts\") pod \"nova-cell1-4cde-account-create-update-9bwcq\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:11 crc kubenswrapper[4758]: I1203 18:33:11.935562 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xbr7\" (UniqueName: \"kubernetes.io/projected/5c270177-5d19-4a1a-a965-f2b106c43e26-kube-api-access-7xbr7\") pod \"nova-cell1-4cde-account-create-update-9bwcq\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:12 crc kubenswrapper[4758]: I1203 18:33:12.023024 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:13 crc kubenswrapper[4758]: W1203 18:33:13.122075 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf27d6cf1_c67d_4b5d_9a54_d1681c949841.slice/crio-9f9f1d459a9f3d7c88d1a2940cd82f8a48351be723100c7b0c68bc900acc2e7c WatchSource:0}: Error finding container 9f9f1d459a9f3d7c88d1a2940cd82f8a48351be723100c7b0c68bc900acc2e7c: Status 404 returned error can't find the container with id 9f9f1d459a9f3d7c88d1a2940cd82f8a48351be723100c7b0c68bc900acc2e7c Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.125487 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bea5-account-create-update-d5xv2"] Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.243625 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cdcf-account-create-update-bftp6"] Dec 03 18:33:13 crc kubenswrapper[4758]: W1203 18:33:13.244992 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2ff9e71_648d_48e2_b428_ee5676abc858.slice/crio-47cd630e505073a4951bc68c4e0660ba49a29585f356a409bd70f0c9cb1b6d62 WatchSource:0}: Error finding container 47cd630e505073a4951bc68c4e0660ba49a29585f356a409bd70f0c9cb1b6d62: Status 404 returned error can't find the container with id 47cd630e505073a4951bc68c4e0660ba49a29585f356a409bd70f0c9cb1b6d62 Dec 03 18:33:13 crc kubenswrapper[4758]: W1203 18:33:13.247243 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod353f3f28_3228_4a1b_9485_8b5f38247264.slice/crio-c7592677d9413d6a1c11a497c1bd9ddc8149286acf1e0776ab83467a89a7e2aa WatchSource:0}: Error finding container c7592677d9413d6a1c11a497c1bd9ddc8149286acf1e0776ab83467a89a7e2aa: Status 404 returned error can't find the container with id c7592677d9413d6a1c11a497c1bd9ddc8149286acf1e0776ab83467a89a7e2aa Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.251047 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9mv8v"] Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.257090 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4cde-account-create-update-9bwcq"] Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.356529 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-knwsq"] Dec 03 18:33:13 crc kubenswrapper[4758]: W1203 18:33:13.361816 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e84029c_4acb_4f87_baee_86fff838a10d.slice/crio-1975a028ac4cdc13356f099947d065c14d7d96422220fc72cb10ae1d83faf3e1 WatchSource:0}: Error finding container 1975a028ac4cdc13356f099947d065c14d7d96422220fc72cb10ae1d83faf3e1: Status 404 returned error can't find the container with id 1975a028ac4cdc13356f099947d065c14d7d96422220fc72cb10ae1d83faf3e1 Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.415584 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-knwsq" event={"ID":"4e84029c-4acb-4f87-baee-86fff838a10d","Type":"ContainerStarted","Data":"1975a028ac4cdc13356f099947d065c14d7d96422220fc72cb10ae1d83faf3e1"} Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.416898 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9mv8v" event={"ID":"c2ff9e71-648d-48e2-b428-ee5676abc858","Type":"ContainerStarted","Data":"47cd630e505073a4951bc68c4e0660ba49a29585f356a409bd70f0c9cb1b6d62"} Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.418111 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bea5-account-create-update-d5xv2" event={"ID":"f27d6cf1-c67d-4b5d-9a54-d1681c949841","Type":"ContainerStarted","Data":"9f9f1d459a9f3d7c88d1a2940cd82f8a48351be723100c7b0c68bc900acc2e7c"} Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.418882 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kct8l"] Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.420165 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" event={"ID":"353f3f28-3228-4a1b-9485-8b5f38247264","Type":"ContainerStarted","Data":"c7592677d9413d6a1c11a497c1bd9ddc8149286acf1e0776ab83467a89a7e2aa"} Dec 03 18:33:13 crc kubenswrapper[4758]: I1203 18:33:13.421830 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" event={"ID":"5c270177-5d19-4a1a-a965-f2b106c43e26","Type":"ContainerStarted","Data":"cf0728adb4b6f61ae155ff59bb0c0e6424681738ae9d6ce920418b2990dd34a3"} Dec 03 18:33:13 crc kubenswrapper[4758]: W1203 18:33:13.425590 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3597b615_57a3_47b9_a718_35e266252642.slice/crio-875c3b3dd5f0cad9e6e373dc493b2812a0b6f2e81c9c5a867cc09af0fe142253 WatchSource:0}: Error finding container 875c3b3dd5f0cad9e6e373dc493b2812a0b6f2e81c9c5a867cc09af0fe142253: Status 404 returned error can't find the container with id 875c3b3dd5f0cad9e6e373dc493b2812a0b6f2e81c9c5a867cc09af0fe142253 Dec 03 18:33:14 crc kubenswrapper[4758]: I1203 18:33:14.442265 4758 generic.go:334] "Generic (PLEG): container finished" podID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerID="b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5" exitCode=0 Dec 03 18:33:14 crc kubenswrapper[4758]: I1203 18:33:14.442362 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmnnm" event={"ID":"f8c3f795-e3d2-452e-a85e-b38746e7abee","Type":"ContainerDied","Data":"b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5"} Dec 03 18:33:14 crc kubenswrapper[4758]: I1203 18:33:14.444448 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bea5-account-create-update-d5xv2" event={"ID":"f27d6cf1-c67d-4b5d-9a54-d1681c949841","Type":"ContainerStarted","Data":"06b63bb05d7325264b426bafee2e2c80bc1d086f57ed26e28882e2b72fe7a416"} Dec 03 18:33:14 crc kubenswrapper[4758]: I1203 18:33:14.446888 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kct8l" event={"ID":"3597b615-57a3-47b9-a718-35e266252642","Type":"ContainerStarted","Data":"875c3b3dd5f0cad9e6e373dc493b2812a0b6f2e81c9c5a867cc09af0fe142253"} Dec 03 18:33:16 crc kubenswrapper[4758]: I1203 18:33:16.469441 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-knwsq" event={"ID":"4e84029c-4acb-4f87-baee-86fff838a10d","Type":"ContainerStarted","Data":"74db9d938180ef9f316a6f59b59b937aa9b0897f0875a115cc87bfdfa642ffce"} Dec 03 18:33:16 crc kubenswrapper[4758]: I1203 18:33:16.471717 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9mv8v" event={"ID":"c2ff9e71-648d-48e2-b428-ee5676abc858","Type":"ContainerStarted","Data":"4807c161f542c4d5869e5dda45e7ea582c7482598d1e385ecf6cf07bef9f5b15"} Dec 03 18:33:16 crc kubenswrapper[4758]: I1203 18:33:16.473322 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kct8l" event={"ID":"3597b615-57a3-47b9-a718-35e266252642","Type":"ContainerStarted","Data":"f414296e923694ce76c9c55c98252f197da7407b1e07f8f474297c09a50a99c9"} Dec 03 18:33:16 crc kubenswrapper[4758]: I1203 18:33:16.474731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" event={"ID":"353f3f28-3228-4a1b-9485-8b5f38247264","Type":"ContainerStarted","Data":"9c22843a689b8b4f13ebc902daacc61dc49234d21a90b2c175cb56c270b849e5"} Dec 03 18:33:16 crc kubenswrapper[4758]: I1203 18:33:16.476428 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" event={"ID":"5c270177-5d19-4a1a-a965-f2b106c43e26","Type":"ContainerStarted","Data":"719ab214c2fe87bdd16e99590cba2e3fe4660c7c3c48eb4a426881a6e76e4dad"} Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.501290 4758 generic.go:334] "Generic (PLEG): container finished" podID="c2ff9e71-648d-48e2-b428-ee5676abc858" containerID="4807c161f542c4d5869e5dda45e7ea582c7482598d1e385ecf6cf07bef9f5b15" exitCode=0 Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.501439 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9mv8v" event={"ID":"c2ff9e71-648d-48e2-b428-ee5676abc858","Type":"ContainerDied","Data":"4807c161f542c4d5869e5dda45e7ea582c7482598d1e385ecf6cf07bef9f5b15"} Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.549755 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-knwsq" podStartSLOduration=7.549735429 podStartE2EDuration="7.549735429s" podCreationTimestamp="2025-12-03 18:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:18.549120192 +0000 UTC m=+5853.750497073" watchObservedRunningTime="2025-12-03 18:33:18.549735429 +0000 UTC m=+5853.751112300" Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.564570 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-bea5-account-create-update-d5xv2" podStartSLOduration=7.564555317 podStartE2EDuration="7.564555317s" podCreationTimestamp="2025-12-03 18:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:18.564106746 +0000 UTC m=+5853.765483607" watchObservedRunningTime="2025-12-03 18:33:18.564555317 +0000 UTC m=+5853.765932178" Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.577146 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" podStartSLOduration=7.577134915 podStartE2EDuration="7.577134915s" podCreationTimestamp="2025-12-03 18:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:18.576216061 +0000 UTC m=+5853.777592962" watchObservedRunningTime="2025-12-03 18:33:18.577134915 +0000 UTC m=+5853.778511776" Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.597432 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-kct8l" podStartSLOduration=7.597409751 podStartE2EDuration="7.597409751s" podCreationTimestamp="2025-12-03 18:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:18.590374102 +0000 UTC m=+5853.791750963" watchObservedRunningTime="2025-12-03 18:33:18.597409751 +0000 UTC m=+5853.798786612" Dec 03 18:33:18 crc kubenswrapper[4758]: I1203 18:33:18.607992 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" podStartSLOduration=7.607974414 podStartE2EDuration="7.607974414s" podCreationTimestamp="2025-12-03 18:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:18.606661069 +0000 UTC m=+5853.808037930" watchObservedRunningTime="2025-12-03 18:33:18.607974414 +0000 UTC m=+5853.809351275" Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.114823 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:33:19 crc kubenswrapper[4758]: E1203 18:33:19.115087 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.510326 4758 generic.go:334] "Generic (PLEG): container finished" podID="3597b615-57a3-47b9-a718-35e266252642" containerID="f414296e923694ce76c9c55c98252f197da7407b1e07f8f474297c09a50a99c9" exitCode=0 Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.510432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kct8l" event={"ID":"3597b615-57a3-47b9-a718-35e266252642","Type":"ContainerDied","Data":"f414296e923694ce76c9c55c98252f197da7407b1e07f8f474297c09a50a99c9"} Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.513710 4758 generic.go:334] "Generic (PLEG): container finished" podID="5c270177-5d19-4a1a-a965-f2b106c43e26" containerID="719ab214c2fe87bdd16e99590cba2e3fe4660c7c3c48eb4a426881a6e76e4dad" exitCode=0 Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.513805 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" event={"ID":"5c270177-5d19-4a1a-a965-f2b106c43e26","Type":"ContainerDied","Data":"719ab214c2fe87bdd16e99590cba2e3fe4660c7c3c48eb4a426881a6e76e4dad"} Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.515915 4758 generic.go:334] "Generic (PLEG): container finished" podID="4e84029c-4acb-4f87-baee-86fff838a10d" containerID="74db9d938180ef9f316a6f59b59b937aa9b0897f0875a115cc87bfdfa642ffce" exitCode=0 Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.515970 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-knwsq" event={"ID":"4e84029c-4acb-4f87-baee-86fff838a10d","Type":"ContainerDied","Data":"74db9d938180ef9f316a6f59b59b937aa9b0897f0875a115cc87bfdfa642ffce"} Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.518797 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmnnm" event={"ID":"f8c3f795-e3d2-452e-a85e-b38746e7abee","Type":"ContainerStarted","Data":"291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a"} Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.520820 4758 generic.go:334] "Generic (PLEG): container finished" podID="f27d6cf1-c67d-4b5d-9a54-d1681c949841" containerID="06b63bb05d7325264b426bafee2e2c80bc1d086f57ed26e28882e2b72fe7a416" exitCode=0 Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.520869 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bea5-account-create-update-d5xv2" event={"ID":"f27d6cf1-c67d-4b5d-9a54-d1681c949841","Type":"ContainerDied","Data":"06b63bb05d7325264b426bafee2e2c80bc1d086f57ed26e28882e2b72fe7a416"} Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.522421 4758 generic.go:334] "Generic (PLEG): container finished" podID="353f3f28-3228-4a1b-9485-8b5f38247264" containerID="9c22843a689b8b4f13ebc902daacc61dc49234d21a90b2c175cb56c270b849e5" exitCode=0 Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.522451 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" event={"ID":"353f3f28-3228-4a1b-9485-8b5f38247264","Type":"ContainerDied","Data":"9c22843a689b8b4f13ebc902daacc61dc49234d21a90b2c175cb56c270b849e5"} Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.591581 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zmnnm" podStartSLOduration=2.636585517 podStartE2EDuration="14.591562373s" podCreationTimestamp="2025-12-03 18:33:05 +0000 UTC" firstStartedPulling="2025-12-03 18:33:06.339297976 +0000 UTC m=+5841.540674837" lastFinishedPulling="2025-12-03 18:33:18.294274842 +0000 UTC m=+5853.495651693" observedRunningTime="2025-12-03 18:33:19.584048112 +0000 UTC m=+5854.785424983" watchObservedRunningTime="2025-12-03 18:33:19.591562373 +0000 UTC m=+5854.792939234" Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.866034 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.975193 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwmmc\" (UniqueName: \"kubernetes.io/projected/c2ff9e71-648d-48e2-b428-ee5676abc858-kube-api-access-dwmmc\") pod \"c2ff9e71-648d-48e2-b428-ee5676abc858\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.975350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ff9e71-648d-48e2-b428-ee5676abc858-operator-scripts\") pod \"c2ff9e71-648d-48e2-b428-ee5676abc858\" (UID: \"c2ff9e71-648d-48e2-b428-ee5676abc858\") " Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.976298 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ff9e71-648d-48e2-b428-ee5676abc858-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2ff9e71-648d-48e2-b428-ee5676abc858" (UID: "c2ff9e71-648d-48e2-b428-ee5676abc858"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.976409 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ff9e71-648d-48e2-b428-ee5676abc858-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:19 crc kubenswrapper[4758]: I1203 18:33:19.982507 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ff9e71-648d-48e2-b428-ee5676abc858-kube-api-access-dwmmc" (OuterVolumeSpecName: "kube-api-access-dwmmc") pod "c2ff9e71-648d-48e2-b428-ee5676abc858" (UID: "c2ff9e71-648d-48e2-b428-ee5676abc858"). InnerVolumeSpecName "kube-api-access-dwmmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.078457 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwmmc\" (UniqueName: \"kubernetes.io/projected/c2ff9e71-648d-48e2-b428-ee5676abc858-kube-api-access-dwmmc\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.532391 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9mv8v" event={"ID":"c2ff9e71-648d-48e2-b428-ee5676abc858","Type":"ContainerDied","Data":"47cd630e505073a4951bc68c4e0660ba49a29585f356a409bd70f0c9cb1b6d62"} Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.532460 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47cd630e505073a4951bc68c4e0660ba49a29585f356a409bd70f0c9cb1b6d62" Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.532545 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9mv8v" Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.876355 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.993766 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27d6cf1-c67d-4b5d-9a54-d1681c949841-operator-scripts\") pod \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.994363 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27d6cf1-c67d-4b5d-9a54-d1681c949841-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f27d6cf1-c67d-4b5d-9a54-d1681c949841" (UID: "f27d6cf1-c67d-4b5d-9a54-d1681c949841"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.995086 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwkpz\" (UniqueName: \"kubernetes.io/projected/f27d6cf1-c67d-4b5d-9a54-d1681c949841-kube-api-access-dwkpz\") pod \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\" (UID: \"f27d6cf1-c67d-4b5d-9a54-d1681c949841\") " Dec 03 18:33:20 crc kubenswrapper[4758]: I1203 18:33:20.997950 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f27d6cf1-c67d-4b5d-9a54-d1681c949841-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.000230 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27d6cf1-c67d-4b5d-9a54-d1681c949841-kube-api-access-dwkpz" (OuterVolumeSpecName: "kube-api-access-dwkpz") pod "f27d6cf1-c67d-4b5d-9a54-d1681c949841" (UID: "f27d6cf1-c67d-4b5d-9a54-d1681c949841"). InnerVolumeSpecName "kube-api-access-dwkpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.077641 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.094008 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.099879 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwkpz\" (UniqueName: \"kubernetes.io/projected/f27d6cf1-c67d-4b5d-9a54-d1681c949841-kube-api-access-dwkpz\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.105430 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.118565 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.200747 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xbr7\" (UniqueName: \"kubernetes.io/projected/5c270177-5d19-4a1a-a965-f2b106c43e26-kube-api-access-7xbr7\") pod \"5c270177-5d19-4a1a-a965-f2b106c43e26\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.200855 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42jzj\" (UniqueName: \"kubernetes.io/projected/4e84029c-4acb-4f87-baee-86fff838a10d-kube-api-access-42jzj\") pod \"4e84029c-4acb-4f87-baee-86fff838a10d\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.200905 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e84029c-4acb-4f87-baee-86fff838a10d-operator-scripts\") pod \"4e84029c-4acb-4f87-baee-86fff838a10d\" (UID: \"4e84029c-4acb-4f87-baee-86fff838a10d\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.200957 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67htp\" (UniqueName: \"kubernetes.io/projected/3597b615-57a3-47b9-a718-35e266252642-kube-api-access-67htp\") pod \"3597b615-57a3-47b9-a718-35e266252642\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.201003 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353f3f28-3228-4a1b-9485-8b5f38247264-operator-scripts\") pod \"353f3f28-3228-4a1b-9485-8b5f38247264\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.201048 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76g85\" (UniqueName: \"kubernetes.io/projected/353f3f28-3228-4a1b-9485-8b5f38247264-kube-api-access-76g85\") pod \"353f3f28-3228-4a1b-9485-8b5f38247264\" (UID: \"353f3f28-3228-4a1b-9485-8b5f38247264\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.201112 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3597b615-57a3-47b9-a718-35e266252642-operator-scripts\") pod \"3597b615-57a3-47b9-a718-35e266252642\" (UID: \"3597b615-57a3-47b9-a718-35e266252642\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.201152 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c270177-5d19-4a1a-a965-f2b106c43e26-operator-scripts\") pod \"5c270177-5d19-4a1a-a965-f2b106c43e26\" (UID: \"5c270177-5d19-4a1a-a965-f2b106c43e26\") " Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.201903 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c270177-5d19-4a1a-a965-f2b106c43e26-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c270177-5d19-4a1a-a965-f2b106c43e26" (UID: "5c270177-5d19-4a1a-a965-f2b106c43e26"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.202180 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e84029c-4acb-4f87-baee-86fff838a10d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4e84029c-4acb-4f87-baee-86fff838a10d" (UID: "4e84029c-4acb-4f87-baee-86fff838a10d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.202192 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3597b615-57a3-47b9-a718-35e266252642-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3597b615-57a3-47b9-a718-35e266252642" (UID: "3597b615-57a3-47b9-a718-35e266252642"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.202181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/353f3f28-3228-4a1b-9485-8b5f38247264-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "353f3f28-3228-4a1b-9485-8b5f38247264" (UID: "353f3f28-3228-4a1b-9485-8b5f38247264"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.204355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e84029c-4acb-4f87-baee-86fff838a10d-kube-api-access-42jzj" (OuterVolumeSpecName: "kube-api-access-42jzj") pod "4e84029c-4acb-4f87-baee-86fff838a10d" (UID: "4e84029c-4acb-4f87-baee-86fff838a10d"). InnerVolumeSpecName "kube-api-access-42jzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.204442 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3597b615-57a3-47b9-a718-35e266252642-kube-api-access-67htp" (OuterVolumeSpecName: "kube-api-access-67htp") pod "3597b615-57a3-47b9-a718-35e266252642" (UID: "3597b615-57a3-47b9-a718-35e266252642"). InnerVolumeSpecName "kube-api-access-67htp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.205731 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353f3f28-3228-4a1b-9485-8b5f38247264-kube-api-access-76g85" (OuterVolumeSpecName: "kube-api-access-76g85") pod "353f3f28-3228-4a1b-9485-8b5f38247264" (UID: "353f3f28-3228-4a1b-9485-8b5f38247264"). InnerVolumeSpecName "kube-api-access-76g85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.206212 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c270177-5d19-4a1a-a965-f2b106c43e26-kube-api-access-7xbr7" (OuterVolumeSpecName: "kube-api-access-7xbr7") pod "5c270177-5d19-4a1a-a965-f2b106c43e26" (UID: "5c270177-5d19-4a1a-a965-f2b106c43e26"). InnerVolumeSpecName "kube-api-access-7xbr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303529 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3597b615-57a3-47b9-a718-35e266252642-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303583 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c270177-5d19-4a1a-a965-f2b106c43e26-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303594 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xbr7\" (UniqueName: \"kubernetes.io/projected/5c270177-5d19-4a1a-a965-f2b106c43e26-kube-api-access-7xbr7\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303606 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42jzj\" (UniqueName: \"kubernetes.io/projected/4e84029c-4acb-4f87-baee-86fff838a10d-kube-api-access-42jzj\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303614 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e84029c-4acb-4f87-baee-86fff838a10d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303622 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67htp\" (UniqueName: \"kubernetes.io/projected/3597b615-57a3-47b9-a718-35e266252642-kube-api-access-67htp\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303631 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353f3f28-3228-4a1b-9485-8b5f38247264-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.303639 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76g85\" (UniqueName: \"kubernetes.io/projected/353f3f28-3228-4a1b-9485-8b5f38247264-kube-api-access-76g85\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.542218 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" event={"ID":"353f3f28-3228-4a1b-9485-8b5f38247264","Type":"ContainerDied","Data":"c7592677d9413d6a1c11a497c1bd9ddc8149286acf1e0776ab83467a89a7e2aa"} Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.542883 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7592677d9413d6a1c11a497c1bd9ddc8149286acf1e0776ab83467a89a7e2aa" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.542280 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdcf-account-create-update-bftp6" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.544100 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kct8l" event={"ID":"3597b615-57a3-47b9-a718-35e266252642","Type":"ContainerDied","Data":"875c3b3dd5f0cad9e6e373dc493b2812a0b6f2e81c9c5a867cc09af0fe142253"} Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.544149 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="875c3b3dd5f0cad9e6e373dc493b2812a0b6f2e81c9c5a867cc09af0fe142253" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.544355 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kct8l" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.545442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" event={"ID":"5c270177-5d19-4a1a-a965-f2b106c43e26","Type":"ContainerDied","Data":"cf0728adb4b6f61ae155ff59bb0c0e6424681738ae9d6ce920418b2990dd34a3"} Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.545490 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf0728adb4b6f61ae155ff59bb0c0e6424681738ae9d6ce920418b2990dd34a3" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.545450 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4cde-account-create-update-9bwcq" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.547336 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-knwsq" event={"ID":"4e84029c-4acb-4f87-baee-86fff838a10d","Type":"ContainerDied","Data":"1975a028ac4cdc13356f099947d065c14d7d96422220fc72cb10ae1d83faf3e1"} Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.547367 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1975a028ac4cdc13356f099947d065c14d7d96422220fc72cb10ae1d83faf3e1" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.547416 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-knwsq" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.548629 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bea5-account-create-update-d5xv2" event={"ID":"f27d6cf1-c67d-4b5d-9a54-d1681c949841","Type":"ContainerDied","Data":"9f9f1d459a9f3d7c88d1a2940cd82f8a48351be723100c7b0c68bc900acc2e7c"} Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.548665 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bea5-account-create-update-d5xv2" Dec 03 18:33:21 crc kubenswrapper[4758]: I1203 18:33:21.548709 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f9f1d459a9f3d7c88d1a2940cd82f8a48351be723100c7b0c68bc900acc2e7c" Dec 03 18:33:25 crc kubenswrapper[4758]: I1203 18:33:25.436101 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:25 crc kubenswrapper[4758]: I1203 18:33:25.437083 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:25 crc kubenswrapper[4758]: I1203 18:33:25.492387 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:25 crc kubenswrapper[4758]: I1203 18:33:25.624919 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:25 crc kubenswrapper[4758]: I1203 18:33:25.730264 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zmnnm"] Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.771858 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wvw82"] Dec 03 18:33:26 crc kubenswrapper[4758]: E1203 18:33:26.772962 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27d6cf1-c67d-4b5d-9a54-d1681c949841" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.773059 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27d6cf1-c67d-4b5d-9a54-d1681c949841" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: E1203 18:33:26.773152 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c270177-5d19-4a1a-a965-f2b106c43e26" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.773227 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c270177-5d19-4a1a-a965-f2b106c43e26" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: E1203 18:33:26.773301 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ff9e71-648d-48e2-b428-ee5676abc858" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.773357 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ff9e71-648d-48e2-b428-ee5676abc858" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: E1203 18:33:26.773422 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3597b615-57a3-47b9-a718-35e266252642" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.773473 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3597b615-57a3-47b9-a718-35e266252642" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: E1203 18:33:26.773546 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353f3f28-3228-4a1b-9485-8b5f38247264" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.773611 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="353f3f28-3228-4a1b-9485-8b5f38247264" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: E1203 18:33:26.773706 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e84029c-4acb-4f87-baee-86fff838a10d" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.773787 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e84029c-4acb-4f87-baee-86fff838a10d" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.774050 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="353f3f28-3228-4a1b-9485-8b5f38247264" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.774142 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c270177-5d19-4a1a-a965-f2b106c43e26" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.774224 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27d6cf1-c67d-4b5d-9a54-d1681c949841" containerName="mariadb-account-create-update" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.774353 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ff9e71-648d-48e2-b428-ee5676abc858" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.774522 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e84029c-4acb-4f87-baee-86fff838a10d" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.774593 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3597b615-57a3-47b9-a718-35e266252642" containerName="mariadb-database-create" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.775397 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.778658 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gd4kw" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.779025 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.779121 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wvw82"] Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.779226 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.928167 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.928239 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-config-data\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.928268 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-scripts\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:26 crc kubenswrapper[4758]: I1203 18:33:26.928345 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg48z\" (UniqueName: \"kubernetes.io/projected/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-kube-api-access-vg48z\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.029634 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg48z\" (UniqueName: \"kubernetes.io/projected/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-kube-api-access-vg48z\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.029761 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.029813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-config-data\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.029846 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-scripts\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.035866 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-config-data\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.035866 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.041054 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-scripts\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.046160 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg48z\" (UniqueName: \"kubernetes.io/projected/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-kube-api-access-vg48z\") pod \"nova-cell0-conductor-db-sync-wvw82\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.100041 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.517774 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wvw82"] Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.602890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wvw82" event={"ID":"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613","Type":"ContainerStarted","Data":"e018f27fedfb5264b4737b21fb9ee22a2c2ac2e00210ade7080dc165dd92c705"} Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.603063 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zmnnm" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="registry-server" containerID="cri-o://291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a" gracePeriod=2 Dec 03 18:33:27 crc kubenswrapper[4758]: I1203 18:33:27.951831 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.049578 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-utilities\") pod \"f8c3f795-e3d2-452e-a85e-b38746e7abee\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.049785 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf7h8\" (UniqueName: \"kubernetes.io/projected/f8c3f795-e3d2-452e-a85e-b38746e7abee-kube-api-access-sf7h8\") pod \"f8c3f795-e3d2-452e-a85e-b38746e7abee\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.049874 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-catalog-content\") pod \"f8c3f795-e3d2-452e-a85e-b38746e7abee\" (UID: \"f8c3f795-e3d2-452e-a85e-b38746e7abee\") " Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.050867 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-utilities" (OuterVolumeSpecName: "utilities") pod "f8c3f795-e3d2-452e-a85e-b38746e7abee" (UID: "f8c3f795-e3d2-452e-a85e-b38746e7abee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.060892 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c3f795-e3d2-452e-a85e-b38746e7abee-kube-api-access-sf7h8" (OuterVolumeSpecName: "kube-api-access-sf7h8") pod "f8c3f795-e3d2-452e-a85e-b38746e7abee" (UID: "f8c3f795-e3d2-452e-a85e-b38746e7abee"). InnerVolumeSpecName "kube-api-access-sf7h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.151982 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.152359 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf7h8\" (UniqueName: \"kubernetes.io/projected/f8c3f795-e3d2-452e-a85e-b38746e7abee-kube-api-access-sf7h8\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.164099 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8c3f795-e3d2-452e-a85e-b38746e7abee" (UID: "f8c3f795-e3d2-452e-a85e-b38746e7abee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.253444 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c3f795-e3d2-452e-a85e-b38746e7abee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.617348 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wvw82" event={"ID":"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613","Type":"ContainerStarted","Data":"49d92bc662262c70244bde1071dd02148d200c1dbcc16958d2146fdf2b7e9500"} Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.621573 4758 generic.go:334] "Generic (PLEG): container finished" podID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerID="291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a" exitCode=0 Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.621616 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmnnm" event={"ID":"f8c3f795-e3d2-452e-a85e-b38746e7abee","Type":"ContainerDied","Data":"291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a"} Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.621639 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zmnnm" event={"ID":"f8c3f795-e3d2-452e-a85e-b38746e7abee","Type":"ContainerDied","Data":"1ed06f7fc8af39a3370d89835e094289c226208d402ebfd9d1ca4a6e86fda7c7"} Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.621658 4758 scope.go:117] "RemoveContainer" containerID="291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.621791 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zmnnm" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.663252 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wvw82" podStartSLOduration=2.663229618 podStartE2EDuration="2.663229618s" podCreationTimestamp="2025-12-03 18:33:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:28.637988259 +0000 UTC m=+5863.839365120" watchObservedRunningTime="2025-12-03 18:33:28.663229618 +0000 UTC m=+5863.864606509" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.667021 4758 scope.go:117] "RemoveContainer" containerID="b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.677031 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zmnnm"] Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.685567 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zmnnm"] Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.697937 4758 scope.go:117] "RemoveContainer" containerID="b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.730847 4758 scope.go:117] "RemoveContainer" containerID="291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a" Dec 03 18:33:28 crc kubenswrapper[4758]: E1203 18:33:28.731265 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a\": container with ID starting with 291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a not found: ID does not exist" containerID="291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.731302 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a"} err="failed to get container status \"291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a\": rpc error: code = NotFound desc = could not find container \"291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a\": container with ID starting with 291c8ebe21042fcb33d22dd45bbe123d3e4b48f84daa73b7ae63d62b6c17d28a not found: ID does not exist" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.731329 4758 scope.go:117] "RemoveContainer" containerID="b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5" Dec 03 18:33:28 crc kubenswrapper[4758]: E1203 18:33:28.731557 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5\": container with ID starting with b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5 not found: ID does not exist" containerID="b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.731580 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5"} err="failed to get container status \"b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5\": rpc error: code = NotFound desc = could not find container \"b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5\": container with ID starting with b475d7d5a5f7d12989c61a09f4235addcf184b7153455c952973cb2a1d2c7fc5 not found: ID does not exist" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.731598 4758 scope.go:117] "RemoveContainer" containerID="b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258" Dec 03 18:33:28 crc kubenswrapper[4758]: E1203 18:33:28.732077 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258\": container with ID starting with b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258 not found: ID does not exist" containerID="b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258" Dec 03 18:33:28 crc kubenswrapper[4758]: I1203 18:33:28.732125 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258"} err="failed to get container status \"b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258\": rpc error: code = NotFound desc = could not find container \"b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258\": container with ID starting with b668996dc4939bf44135d80ade9b0f950b020726ce9fb5cc7e2875c0d7362258 not found: ID does not exist" Dec 03 18:33:29 crc kubenswrapper[4758]: I1203 18:33:29.127749 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" path="/var/lib/kubelet/pods/f8c3f795-e3d2-452e-a85e-b38746e7abee/volumes" Dec 03 18:33:34 crc kubenswrapper[4758]: I1203 18:33:34.114142 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:33:34 crc kubenswrapper[4758]: E1203 18:33:34.114943 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:33:34 crc kubenswrapper[4758]: I1203 18:33:34.679234 4758 generic.go:334] "Generic (PLEG): container finished" podID="b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" containerID="49d92bc662262c70244bde1071dd02148d200c1dbcc16958d2146fdf2b7e9500" exitCode=0 Dec 03 18:33:34 crc kubenswrapper[4758]: I1203 18:33:34.679333 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wvw82" event={"ID":"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613","Type":"ContainerDied","Data":"49d92bc662262c70244bde1071dd02148d200c1dbcc16958d2146fdf2b7e9500"} Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.009352 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.090831 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg48z\" (UniqueName: \"kubernetes.io/projected/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-kube-api-access-vg48z\") pod \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.090920 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-scripts\") pod \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.091052 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-combined-ca-bundle\") pod \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.091083 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-config-data\") pod \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\" (UID: \"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613\") " Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.096321 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-kube-api-access-vg48z" (OuterVolumeSpecName: "kube-api-access-vg48z") pod "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" (UID: "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613"). InnerVolumeSpecName "kube-api-access-vg48z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.096598 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-scripts" (OuterVolumeSpecName: "scripts") pod "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" (UID: "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.117343 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" (UID: "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.119362 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-config-data" (OuterVolumeSpecName: "config-data") pod "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" (UID: "b9f5e5bc-c00d-4cbf-90af-0d58b19eb613"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.193219 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg48z\" (UniqueName: \"kubernetes.io/projected/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-kube-api-access-vg48z\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.193271 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.193284 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.193295 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.697943 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wvw82" event={"ID":"b9f5e5bc-c00d-4cbf-90af-0d58b19eb613","Type":"ContainerDied","Data":"e018f27fedfb5264b4737b21fb9ee22a2c2ac2e00210ade7080dc165dd92c705"} Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.697981 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e018f27fedfb5264b4737b21fb9ee22a2c2ac2e00210ade7080dc165dd92c705" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.698013 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wvw82" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.782584 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:33:36 crc kubenswrapper[4758]: E1203 18:33:36.782989 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="extract-content" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.783005 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="extract-content" Dec 03 18:33:36 crc kubenswrapper[4758]: E1203 18:33:36.783019 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="extract-utilities" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.783026 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="extract-utilities" Dec 03 18:33:36 crc kubenswrapper[4758]: E1203 18:33:36.783046 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" containerName="nova-cell0-conductor-db-sync" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.783053 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" containerName="nova-cell0-conductor-db-sync" Dec 03 18:33:36 crc kubenswrapper[4758]: E1203 18:33:36.783069 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="registry-server" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.783075 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="registry-server" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.783239 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" containerName="nova-cell0-conductor-db-sync" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.783254 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c3f795-e3d2-452e-a85e-b38746e7abee" containerName="registry-server" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.784112 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.787454 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gd4kw" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.795442 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.804761 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.907991 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wznb5\" (UniqueName: \"kubernetes.io/projected/39670374-d4c0-40d2-b1c2-d838b472fdc1-kube-api-access-wznb5\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.908357 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:36 crc kubenswrapper[4758]: I1203 18:33:36.908401 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.010320 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.010760 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wznb5\" (UniqueName: \"kubernetes.io/projected/39670374-d4c0-40d2-b1c2-d838b472fdc1-kube-api-access-wznb5\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.010884 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.014603 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.018151 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.041313 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wznb5\" (UniqueName: \"kubernetes.io/projected/39670374-d4c0-40d2-b1c2-d838b472fdc1-kube-api-access-wznb5\") pod \"nova-cell0-conductor-0\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.102023 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.565614 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:33:37 crc kubenswrapper[4758]: I1203 18:33:37.709721 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"39670374-d4c0-40d2-b1c2-d838b472fdc1","Type":"ContainerStarted","Data":"024ca130aaca69ce7f856fc67c3ed4b89f85b181529be99e2c8b734d695f0450"} Dec 03 18:33:38 crc kubenswrapper[4758]: I1203 18:33:38.722641 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"39670374-d4c0-40d2-b1c2-d838b472fdc1","Type":"ContainerStarted","Data":"ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf"} Dec 03 18:33:38 crc kubenswrapper[4758]: I1203 18:33:38.723316 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:38 crc kubenswrapper[4758]: I1203 18:33:38.752187 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.752156727 podStartE2EDuration="2.752156727s" podCreationTimestamp="2025-12-03 18:33:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:38.73779193 +0000 UTC m=+5873.939168861" watchObservedRunningTime="2025-12-03 18:33:38.752156727 +0000 UTC m=+5873.953533598" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.129363 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.566184 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-wlql4"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.567856 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.570259 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.570382 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.586487 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlql4"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.678785 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.687115 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.694378 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.714464 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-scripts\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.714534 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-config-data\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.714699 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.719613 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lttv\" (UniqueName: \"kubernetes.io/projected/a87223ea-1a03-43c6-bbe0-e9a11b4db504-kube-api-access-2lttv\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.738759 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.768503 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.771218 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.778226 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.795029 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831061 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-config-data\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lttv\" (UniqueName: \"kubernetes.io/projected/a87223ea-1a03-43c6-bbe0-e9a11b4db504-kube-api-access-2lttv\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831164 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-scripts\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-config-data\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831254 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqjv\" (UniqueName: \"kubernetes.io/projected/2828e8f9-352b-4718-990e-e9db59f6ba4a-kube-api-access-mgqjv\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831315 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.831732 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.833323 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.841374 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.846773 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.848931 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-scripts\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.860859 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-config-data\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.862536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.872528 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lttv\" (UniqueName: \"kubernetes.io/projected/a87223ea-1a03-43c6-bbe0-e9a11b4db504-kube-api-access-2lttv\") pod \"nova-cell0-cell-mapping-wlql4\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.876417 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.877923 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.880651 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.884829 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.889850 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.932573 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr4pw\" (UniqueName: \"kubernetes.io/projected/bfa37d80-1e92-4a29-82f2-66f902c2021c-kube-api-access-jr4pw\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.932630 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.933183 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.933267 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-config-data\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.933330 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxdj2\" (UniqueName: \"kubernetes.io/projected/f4e3e6b7-8244-48b9-9947-2547938d5206-kube-api-access-kxdj2\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.933373 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.933434 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.933570 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-config-data\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.934621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqjv\" (UniqueName: \"kubernetes.io/projected/2828e8f9-352b-4718-990e-e9db59f6ba4a-kube-api-access-mgqjv\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.934661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfa37d80-1e92-4a29-82f2-66f902c2021c-logs\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.937439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-config-data\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.938656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.961218 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqjv\" (UniqueName: \"kubernetes.io/projected/2828e8f9-352b-4718-990e-e9db59f6ba4a-kube-api-access-mgqjv\") pod \"nova-scheduler-0\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:42 crc kubenswrapper[4758]: I1203 18:33:42.993881 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7647579f87-n8tn8"] Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.015221 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.026329 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7647579f87-n8tn8"] Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.026431 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.035865 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slvh5\" (UniqueName: \"kubernetes.io/projected/3b652247-e8c0-4986-9d98-156703326dee-kube-api-access-slvh5\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.035941 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxdj2\" (UniqueName: \"kubernetes.io/projected/f4e3e6b7-8244-48b9-9947-2547938d5206-kube-api-access-kxdj2\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.035982 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036025 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-config-data\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036073 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfa37d80-1e92-4a29-82f2-66f902c2021c-logs\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036110 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr4pw\" (UniqueName: \"kubernetes.io/projected/bfa37d80-1e92-4a29-82f2-66f902c2021c-kube-api-access-jr4pw\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036132 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036169 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-config-data\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036189 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036212 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b652247-e8c0-4986-9d98-156703326dee-logs\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.036234 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.050937 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.051609 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfa37d80-1e92-4a29-82f2-66f902c2021c-logs\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.056384 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.056981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-config-data\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.059621 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.068973 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr4pw\" (UniqueName: \"kubernetes.io/projected/bfa37d80-1e92-4a29-82f2-66f902c2021c-kube-api-access-jr4pw\") pod \"nova-api-0\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.071583 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxdj2\" (UniqueName: \"kubernetes.io/projected/f4e3e6b7-8244-48b9-9947-2547938d5206-kube-api-access-kxdj2\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.098844 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.137543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b652247-e8c0-4986-9d98-156703326dee-logs\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.137592 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-sb\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.137651 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slvh5\" (UniqueName: \"kubernetes.io/projected/3b652247-e8c0-4986-9d98-156703326dee-kube-api-access-slvh5\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.137972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b652247-e8c0-4986-9d98-156703326dee-logs\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.138036 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-nb\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.138084 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g7p6\" (UniqueName: \"kubernetes.io/projected/21a13077-0749-4821-8735-06d720d6079c-kube-api-access-5g7p6\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.138106 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-dns-svc\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.138128 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-config\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.138175 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-config-data\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.138190 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.143608 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-config-data\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.148514 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.154248 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slvh5\" (UniqueName: \"kubernetes.io/projected/3b652247-e8c0-4986-9d98-156703326dee-kube-api-access-slvh5\") pod \"nova-metadata-0\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.244552 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g7p6\" (UniqueName: \"kubernetes.io/projected/21a13077-0749-4821-8735-06d720d6079c-kube-api-access-5g7p6\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.245001 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-dns-svc\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.246672 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-dns-svc\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.248482 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-config\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.250095 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-sb\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.250528 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-nb\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.257724 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-config\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.257895 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-nb\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.258452 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-sb\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.262837 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g7p6\" (UniqueName: \"kubernetes.io/projected/21a13077-0749-4821-8735-06d720d6079c-kube-api-access-5g7p6\") pod \"dnsmasq-dns-7647579f87-n8tn8\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.331453 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.374093 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.389568 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.390004 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlql4"] Dec 03 18:33:43 crc kubenswrapper[4758]: W1203 18:33:43.402935 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87223ea_1a03_43c6_bbe0_e9a11b4db504.slice/crio-db5d4791148fc5ab7f56f3b46f3c267f3b372cb3b1fac2e6c1b09b4ddb85a61d WatchSource:0}: Error finding container db5d4791148fc5ab7f56f3b46f3c267f3b372cb3b1fac2e6c1b09b4ddb85a61d: Status 404 returned error can't find the container with id db5d4791148fc5ab7f56f3b46f3c267f3b372cb3b1fac2e6c1b09b4ddb85a61d Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.403009 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.474862 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h6pbk"] Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.476056 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.480005 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.480077 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.488846 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h6pbk"] Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.657582 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-config-data\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.658083 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.658136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85fsq\" (UniqueName: \"kubernetes.io/projected/8a5dd0a8-1259-4bca-b035-485c40718803-kube-api-access-85fsq\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.658179 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-scripts\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.717359 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:43 crc kubenswrapper[4758]: W1203 18:33:43.720031 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfa37d80_1e92_4a29_82f2_66f902c2021c.slice/crio-2a73f7a798761b0ca942832aca579e315d5f0da4431834d5c6c41fd8724a1ff5 WatchSource:0}: Error finding container 2a73f7a798761b0ca942832aca579e315d5f0da4431834d5c6c41fd8724a1ff5: Status 404 returned error can't find the container with id 2a73f7a798761b0ca942832aca579e315d5f0da4431834d5c6c41fd8724a1ff5 Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.760002 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.760070 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85fsq\" (UniqueName: \"kubernetes.io/projected/8a5dd0a8-1259-4bca-b035-485c40718803-kube-api-access-85fsq\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.760119 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-scripts\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.760160 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-config-data\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.769317 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-config-data\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.772350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-scripts\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: W1203 18:33:43.772900 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4e3e6b7_8244_48b9_9947_2547938d5206.slice/crio-476de057ee896558bbcfe44097869e9c14f45cdf71895de2b7e75d812d559a64 WatchSource:0}: Error finding container 476de057ee896558bbcfe44097869e9c14f45cdf71895de2b7e75d812d559a64: Status 404 returned error can't find the container with id 476de057ee896558bbcfe44097869e9c14f45cdf71895de2b7e75d812d559a64 Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.773136 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.773940 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.778266 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85fsq\" (UniqueName: \"kubernetes.io/projected/8a5dd0a8-1259-4bca-b035-485c40718803-kube-api-access-85fsq\") pod \"nova-cell1-conductor-db-sync-h6pbk\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.778446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2828e8f9-352b-4718-990e-e9db59f6ba4a","Type":"ContainerStarted","Data":"ca032caaf97420fcdb37518e602b1baa1e546a683baffa70667096665d842291"} Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.779959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bfa37d80-1e92-4a29-82f2-66f902c2021c","Type":"ContainerStarted","Data":"2a73f7a798761b0ca942832aca579e315d5f0da4431834d5c6c41fd8724a1ff5"} Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.781070 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlql4" event={"ID":"a87223ea-1a03-43c6-bbe0-e9a11b4db504","Type":"ContainerStarted","Data":"db5d4791148fc5ab7f56f3b46f3c267f3b372cb3b1fac2e6c1b09b4ddb85a61d"} Dec 03 18:33:43 crc kubenswrapper[4758]: I1203 18:33:43.808420 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.042517 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7647579f87-n8tn8"] Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.070278 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.077752 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h6pbk"] Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.793229 4758 generic.go:334] "Generic (PLEG): container finished" podID="21a13077-0749-4821-8735-06d720d6079c" containerID="8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0" exitCode=0 Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.793278 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" event={"ID":"21a13077-0749-4821-8735-06d720d6079c","Type":"ContainerDied","Data":"8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.793784 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" event={"ID":"21a13077-0749-4821-8735-06d720d6079c","Type":"ContainerStarted","Data":"d7c36ee6c34a2974a42f53cb3363165a2ec58fbc4a124da99930f00ff99d174d"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.798186 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" event={"ID":"8a5dd0a8-1259-4bca-b035-485c40718803","Type":"ContainerStarted","Data":"4d71f0cfcdd494e80f178e79e515c74c807b82103e6484afeedb305b61a28287"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.798230 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" event={"ID":"8a5dd0a8-1259-4bca-b035-485c40718803","Type":"ContainerStarted","Data":"9f0603ff98734b7ac98f6aa49de9ea24cde63f63881dcdcf55da157e90f826d1"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.800090 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2828e8f9-352b-4718-990e-e9db59f6ba4a","Type":"ContainerStarted","Data":"bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.802177 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bfa37d80-1e92-4a29-82f2-66f902c2021c","Type":"ContainerStarted","Data":"bd02c354adfd86efb5b82889faf36bec8dac63cdf5fa6ddba2ef2b5c75c9a449"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.804540 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b652247-e8c0-4986-9d98-156703326dee","Type":"ContainerStarted","Data":"4beb4de623e45e953a79c9bf05163519762a9ce4a1d730f676284f6a1b5d3519"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.804592 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b652247-e8c0-4986-9d98-156703326dee","Type":"ContainerStarted","Data":"58995773b2c9f34f86487a366700fead2864f46948aea80075ea5849cdbf5ae0"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.807611 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f4e3e6b7-8244-48b9-9947-2547938d5206","Type":"ContainerStarted","Data":"cfb70d3b0fe2005455a60707770bdb91f5e52dd5a9ac0c7c57b2670fcc375fcc"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.807661 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f4e3e6b7-8244-48b9-9947-2547938d5206","Type":"ContainerStarted","Data":"476de057ee896558bbcfe44097869e9c14f45cdf71895de2b7e75d812d559a64"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.814954 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlql4" event={"ID":"a87223ea-1a03-43c6-bbe0-e9a11b4db504","Type":"ContainerStarted","Data":"0a575e5ae3d1e7d4609a1146f1e04466c6e88ef47d66f4fd4e4f690ba0aeadbd"} Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.835994 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.835974369 podStartE2EDuration="2.835974369s" podCreationTimestamp="2025-12-03 18:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:44.835919607 +0000 UTC m=+5880.037296488" watchObservedRunningTime="2025-12-03 18:33:44.835974369 +0000 UTC m=+5880.037351230" Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.857046 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.857029425 podStartE2EDuration="2.857029425s" podCreationTimestamp="2025-12-03 18:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:44.850740445 +0000 UTC m=+5880.052117336" watchObservedRunningTime="2025-12-03 18:33:44.857029425 +0000 UTC m=+5880.058406286" Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.883900 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" podStartSLOduration=1.883880456 podStartE2EDuration="1.883880456s" podCreationTimestamp="2025-12-03 18:33:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:44.866661484 +0000 UTC m=+5880.068038385" watchObservedRunningTime="2025-12-03 18:33:44.883880456 +0000 UTC m=+5880.085257317" Dec 03 18:33:44 crc kubenswrapper[4758]: I1203 18:33:44.886503 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-wlql4" podStartSLOduration=2.886495097 podStartE2EDuration="2.886495097s" podCreationTimestamp="2025-12-03 18:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:44.884045741 +0000 UTC m=+5880.085422612" watchObservedRunningTime="2025-12-03 18:33:44.886495097 +0000 UTC m=+5880.087871958" Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.823539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bfa37d80-1e92-4a29-82f2-66f902c2021c","Type":"ContainerStarted","Data":"10dffd0c3394323fc7bcc1d00c465657f5ce32931dbdb98fa35ca149180eec75"} Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.827649 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b652247-e8c0-4986-9d98-156703326dee","Type":"ContainerStarted","Data":"544eb92443af687a456ca0efc9fa52b4ad48c28f44e325f466f149b0fea393b2"} Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.830727 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" event={"ID":"21a13077-0749-4821-8735-06d720d6079c","Type":"ContainerStarted","Data":"07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0"} Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.831177 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.842600 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.842579566 podStartE2EDuration="3.842579566s" podCreationTimestamp="2025-12-03 18:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:45.840899371 +0000 UTC m=+5881.042276272" watchObservedRunningTime="2025-12-03 18:33:45.842579566 +0000 UTC m=+5881.043956427" Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.873226 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.873204339 podStartE2EDuration="3.873204339s" podCreationTimestamp="2025-12-03 18:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:45.858822262 +0000 UTC m=+5881.060199133" watchObservedRunningTime="2025-12-03 18:33:45.873204339 +0000 UTC m=+5881.074581200" Dec 03 18:33:45 crc kubenswrapper[4758]: I1203 18:33:45.886902 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" podStartSLOduration=3.886884496 podStartE2EDuration="3.886884496s" podCreationTimestamp="2025-12-03 18:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:45.881019769 +0000 UTC m=+5881.082396640" watchObservedRunningTime="2025-12-03 18:33:45.886884496 +0000 UTC m=+5881.088261357" Dec 03 18:33:46 crc kubenswrapper[4758]: I1203 18:33:46.115195 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:33:46 crc kubenswrapper[4758]: E1203 18:33:46.115865 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:33:48 crc kubenswrapper[4758]: I1203 18:33:48.017608 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 18:33:48 crc kubenswrapper[4758]: I1203 18:33:48.374768 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:48 crc kubenswrapper[4758]: I1203 18:33:48.390119 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:33:48 crc kubenswrapper[4758]: I1203 18:33:48.390166 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:33:48 crc kubenswrapper[4758]: I1203 18:33:48.877273 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a5dd0a8-1259-4bca-b035-485c40718803" containerID="4d71f0cfcdd494e80f178e79e515c74c807b82103e6484afeedb305b61a28287" exitCode=0 Dec 03 18:33:48 crc kubenswrapper[4758]: I1203 18:33:48.877315 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" event={"ID":"8a5dd0a8-1259-4bca-b035-485c40718803","Type":"ContainerDied","Data":"4d71f0cfcdd494e80f178e79e515c74c807b82103e6484afeedb305b61a28287"} Dec 03 18:33:49 crc kubenswrapper[4758]: I1203 18:33:49.888386 4758 generic.go:334] "Generic (PLEG): container finished" podID="a87223ea-1a03-43c6-bbe0-e9a11b4db504" containerID="0a575e5ae3d1e7d4609a1146f1e04466c6e88ef47d66f4fd4e4f690ba0aeadbd" exitCode=0 Dec 03 18:33:49 crc kubenswrapper[4758]: I1203 18:33:49.888467 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlql4" event={"ID":"a87223ea-1a03-43c6-bbe0-e9a11b4db504","Type":"ContainerDied","Data":"0a575e5ae3d1e7d4609a1146f1e04466c6e88ef47d66f4fd4e4f690ba0aeadbd"} Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.268433 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.333666 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-combined-ca-bundle\") pod \"8a5dd0a8-1259-4bca-b035-485c40718803\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.333735 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-scripts\") pod \"8a5dd0a8-1259-4bca-b035-485c40718803\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.333772 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85fsq\" (UniqueName: \"kubernetes.io/projected/8a5dd0a8-1259-4bca-b035-485c40718803-kube-api-access-85fsq\") pod \"8a5dd0a8-1259-4bca-b035-485c40718803\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.333809 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-config-data\") pod \"8a5dd0a8-1259-4bca-b035-485c40718803\" (UID: \"8a5dd0a8-1259-4bca-b035-485c40718803\") " Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.339657 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-scripts" (OuterVolumeSpecName: "scripts") pod "8a5dd0a8-1259-4bca-b035-485c40718803" (UID: "8a5dd0a8-1259-4bca-b035-485c40718803"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.340588 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5dd0a8-1259-4bca-b035-485c40718803-kube-api-access-85fsq" (OuterVolumeSpecName: "kube-api-access-85fsq") pod "8a5dd0a8-1259-4bca-b035-485c40718803" (UID: "8a5dd0a8-1259-4bca-b035-485c40718803"). InnerVolumeSpecName "kube-api-access-85fsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.360030 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-config-data" (OuterVolumeSpecName: "config-data") pod "8a5dd0a8-1259-4bca-b035-485c40718803" (UID: "8a5dd0a8-1259-4bca-b035-485c40718803"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.366873 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a5dd0a8-1259-4bca-b035-485c40718803" (UID: "8a5dd0a8-1259-4bca-b035-485c40718803"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.436027 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.436060 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.436069 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85fsq\" (UniqueName: \"kubernetes.io/projected/8a5dd0a8-1259-4bca-b035-485c40718803-kube-api-access-85fsq\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.436078 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a5dd0a8-1259-4bca-b035-485c40718803-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.898619 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" event={"ID":"8a5dd0a8-1259-4bca-b035-485c40718803","Type":"ContainerDied","Data":"9f0603ff98734b7ac98f6aa49de9ea24cde63f63881dcdcf55da157e90f826d1"} Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.898653 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h6pbk" Dec 03 18:33:50 crc kubenswrapper[4758]: I1203 18:33:50.898668 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f0603ff98734b7ac98f6aa49de9ea24cde63f63881dcdcf55da157e90f826d1" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.010974 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:33:51 crc kubenswrapper[4758]: E1203 18:33:51.011467 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5dd0a8-1259-4bca-b035-485c40718803" containerName="nova-cell1-conductor-db-sync" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.011487 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5dd0a8-1259-4bca-b035-485c40718803" containerName="nova-cell1-conductor-db-sync" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.011746 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5dd0a8-1259-4bca-b035-485c40718803" containerName="nova-cell1-conductor-db-sync" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.012438 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.016759 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.032563 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.048502 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.048737 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnhgf\" (UniqueName: \"kubernetes.io/projected/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-kube-api-access-qnhgf\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.048815 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.150149 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnhgf\" (UniqueName: \"kubernetes.io/projected/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-kube-api-access-qnhgf\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.150244 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.150278 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.155488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.155563 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.166413 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnhgf\" (UniqueName: \"kubernetes.io/projected/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-kube-api-access-qnhgf\") pod \"nova-cell1-conductor-0\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.225877 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.330483 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.355389 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lttv\" (UniqueName: \"kubernetes.io/projected/a87223ea-1a03-43c6-bbe0-e9a11b4db504-kube-api-access-2lttv\") pod \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.355486 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-config-data\") pod \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.355533 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-combined-ca-bundle\") pod \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.355665 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-scripts\") pod \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\" (UID: \"a87223ea-1a03-43c6-bbe0-e9a11b4db504\") " Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.359348 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87223ea-1a03-43c6-bbe0-e9a11b4db504-kube-api-access-2lttv" (OuterVolumeSpecName: "kube-api-access-2lttv") pod "a87223ea-1a03-43c6-bbe0-e9a11b4db504" (UID: "a87223ea-1a03-43c6-bbe0-e9a11b4db504"). InnerVolumeSpecName "kube-api-access-2lttv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.360232 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-scripts" (OuterVolumeSpecName: "scripts") pod "a87223ea-1a03-43c6-bbe0-e9a11b4db504" (UID: "a87223ea-1a03-43c6-bbe0-e9a11b4db504"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.389913 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a87223ea-1a03-43c6-bbe0-e9a11b4db504" (UID: "a87223ea-1a03-43c6-bbe0-e9a11b4db504"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.392550 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-config-data" (OuterVolumeSpecName: "config-data") pod "a87223ea-1a03-43c6-bbe0-e9a11b4db504" (UID: "a87223ea-1a03-43c6-bbe0-e9a11b4db504"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.459434 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.459483 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.459503 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lttv\" (UniqueName: \"kubernetes.io/projected/a87223ea-1a03-43c6-bbe0-e9a11b4db504-kube-api-access-2lttv\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.459517 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a87223ea-1a03-43c6-bbe0-e9a11b4db504-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.773977 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:33:51 crc kubenswrapper[4758]: W1203 18:33:51.778660 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e03e4bb_9e94_4ff1_bc70_e74cfc176094.slice/crio-3141891fd34bb66c7ae90cffbcb9df9c50c7807263610a09a0929e119674a427 WatchSource:0}: Error finding container 3141891fd34bb66c7ae90cffbcb9df9c50c7807263610a09a0929e119674a427: Status 404 returned error can't find the container with id 3141891fd34bb66c7ae90cffbcb9df9c50c7807263610a09a0929e119674a427 Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.908499 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wlql4" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.908507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wlql4" event={"ID":"a87223ea-1a03-43c6-bbe0-e9a11b4db504","Type":"ContainerDied","Data":"db5d4791148fc5ab7f56f3b46f3c267f3b372cb3b1fac2e6c1b09b4ddb85a61d"} Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.908966 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db5d4791148fc5ab7f56f3b46f3c267f3b372cb3b1fac2e6c1b09b4ddb85a61d" Dec 03 18:33:51 crc kubenswrapper[4758]: I1203 18:33:51.910075 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4e03e4bb-9e94-4ff1-bc70-e74cfc176094","Type":"ContainerStarted","Data":"3141891fd34bb66c7ae90cffbcb9df9c50c7807263610a09a0929e119674a427"} Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.089510 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.089874 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-log" containerID="cri-o://bd02c354adfd86efb5b82889faf36bec8dac63cdf5fa6ddba2ef2b5c75c9a449" gracePeriod=30 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.089950 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-api" containerID="cri-o://10dffd0c3394323fc7bcc1d00c465657f5ce32931dbdb98fa35ca149180eec75" gracePeriod=30 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.172107 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.172433 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2828e8f9-352b-4718-990e-e9db59f6ba4a" containerName="nova-scheduler-scheduler" containerID="cri-o://bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f" gracePeriod=30 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.188098 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.188351 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-log" containerID="cri-o://4beb4de623e45e953a79c9bf05163519762a9ce4a1d730f676284f6a1b5d3519" gracePeriod=30 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.188877 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-metadata" containerID="cri-o://544eb92443af687a456ca0efc9fa52b4ad48c28f44e325f466f149b0fea393b2" gracePeriod=30 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.924801 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4e03e4bb-9e94-4ff1-bc70-e74cfc176094","Type":"ContainerStarted","Data":"3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84"} Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.926058 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.934616 4758 generic.go:334] "Generic (PLEG): container finished" podID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerID="10dffd0c3394323fc7bcc1d00c465657f5ce32931dbdb98fa35ca149180eec75" exitCode=0 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.934642 4758 generic.go:334] "Generic (PLEG): container finished" podID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerID="bd02c354adfd86efb5b82889faf36bec8dac63cdf5fa6ddba2ef2b5c75c9a449" exitCode=143 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.934728 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bfa37d80-1e92-4a29-82f2-66f902c2021c","Type":"ContainerDied","Data":"10dffd0c3394323fc7bcc1d00c465657f5ce32931dbdb98fa35ca149180eec75"} Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.934760 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bfa37d80-1e92-4a29-82f2-66f902c2021c","Type":"ContainerDied","Data":"bd02c354adfd86efb5b82889faf36bec8dac63cdf5fa6ddba2ef2b5c75c9a449"} Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.937309 4758 generic.go:334] "Generic (PLEG): container finished" podID="3b652247-e8c0-4986-9d98-156703326dee" containerID="544eb92443af687a456ca0efc9fa52b4ad48c28f44e325f466f149b0fea393b2" exitCode=0 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.937332 4758 generic.go:334] "Generic (PLEG): container finished" podID="3b652247-e8c0-4986-9d98-156703326dee" containerID="4beb4de623e45e953a79c9bf05163519762a9ce4a1d730f676284f6a1b5d3519" exitCode=143 Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.937355 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b652247-e8c0-4986-9d98-156703326dee","Type":"ContainerDied","Data":"544eb92443af687a456ca0efc9fa52b4ad48c28f44e325f466f149b0fea393b2"} Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.937374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b652247-e8c0-4986-9d98-156703326dee","Type":"ContainerDied","Data":"4beb4de623e45e953a79c9bf05163519762a9ce4a1d730f676284f6a1b5d3519"} Dec 03 18:33:52 crc kubenswrapper[4758]: I1203 18:33:52.959761 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.959739933 podStartE2EDuration="2.959739933s" podCreationTimestamp="2025-12-03 18:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:52.944526944 +0000 UTC m=+5888.145903805" watchObservedRunningTime="2025-12-03 18:33:52.959739933 +0000 UTC m=+5888.161116794" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.133059 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.295225 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.296077 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr4pw\" (UniqueName: \"kubernetes.io/projected/bfa37d80-1e92-4a29-82f2-66f902c2021c-kube-api-access-jr4pw\") pod \"bfa37d80-1e92-4a29-82f2-66f902c2021c\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.296148 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-combined-ca-bundle\") pod \"bfa37d80-1e92-4a29-82f2-66f902c2021c\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.296256 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfa37d80-1e92-4a29-82f2-66f902c2021c-logs\") pod \"bfa37d80-1e92-4a29-82f2-66f902c2021c\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.296326 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-config-data\") pod \"bfa37d80-1e92-4a29-82f2-66f902c2021c\" (UID: \"bfa37d80-1e92-4a29-82f2-66f902c2021c\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.296624 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa37d80-1e92-4a29-82f2-66f902c2021c-logs" (OuterVolumeSpecName: "logs") pod "bfa37d80-1e92-4a29-82f2-66f902c2021c" (UID: "bfa37d80-1e92-4a29-82f2-66f902c2021c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.297155 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfa37d80-1e92-4a29-82f2-66f902c2021c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.301473 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa37d80-1e92-4a29-82f2-66f902c2021c-kube-api-access-jr4pw" (OuterVolumeSpecName: "kube-api-access-jr4pw") pod "bfa37d80-1e92-4a29-82f2-66f902c2021c" (UID: "bfa37d80-1e92-4a29-82f2-66f902c2021c"). InnerVolumeSpecName "kube-api-access-jr4pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.324982 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-config-data" (OuterVolumeSpecName: "config-data") pod "bfa37d80-1e92-4a29-82f2-66f902c2021c" (UID: "bfa37d80-1e92-4a29-82f2-66f902c2021c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.338971 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfa37d80-1e92-4a29-82f2-66f902c2021c" (UID: "bfa37d80-1e92-4a29-82f2-66f902c2021c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.375117 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.388887 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.398558 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slvh5\" (UniqueName: \"kubernetes.io/projected/3b652247-e8c0-4986-9d98-156703326dee-kube-api-access-slvh5\") pod \"3b652247-e8c0-4986-9d98-156703326dee\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.398819 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-combined-ca-bundle\") pod \"3b652247-e8c0-4986-9d98-156703326dee\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.398960 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-config-data\") pod \"3b652247-e8c0-4986-9d98-156703326dee\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.399339 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b652247-e8c0-4986-9d98-156703326dee-logs\") pod \"3b652247-e8c0-4986-9d98-156703326dee\" (UID: \"3b652247-e8c0-4986-9d98-156703326dee\") " Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.400304 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.400448 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr4pw\" (UniqueName: \"kubernetes.io/projected/bfa37d80-1e92-4a29-82f2-66f902c2021c-kube-api-access-jr4pw\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.400619 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa37d80-1e92-4a29-82f2-66f902c2021c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.401334 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b652247-e8c0-4986-9d98-156703326dee-logs" (OuterVolumeSpecName: "logs") pod "3b652247-e8c0-4986-9d98-156703326dee" (UID: "3b652247-e8c0-4986-9d98-156703326dee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.401623 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b652247-e8c0-4986-9d98-156703326dee-kube-api-access-slvh5" (OuterVolumeSpecName: "kube-api-access-slvh5") pod "3b652247-e8c0-4986-9d98-156703326dee" (UID: "3b652247-e8c0-4986-9d98-156703326dee"). InnerVolumeSpecName "kube-api-access-slvh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.404913 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.432113 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b652247-e8c0-4986-9d98-156703326dee" (UID: "3b652247-e8c0-4986-9d98-156703326dee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.441275 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-config-data" (OuterVolumeSpecName: "config-data") pod "3b652247-e8c0-4986-9d98-156703326dee" (UID: "3b652247-e8c0-4986-9d98-156703326dee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.468182 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5584c87ddf-q6bzc"] Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.502591 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b652247-e8c0-4986-9d98-156703326dee-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.502620 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slvh5\" (UniqueName: \"kubernetes.io/projected/3b652247-e8c0-4986-9d98-156703326dee-kube-api-access-slvh5\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.502628 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.502637 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b652247-e8c0-4986-9d98-156703326dee-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.949287 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.949299 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bfa37d80-1e92-4a29-82f2-66f902c2021c","Type":"ContainerDied","Data":"2a73f7a798761b0ca942832aca579e315d5f0da4431834d5c6c41fd8724a1ff5"} Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.950256 4758 scope.go:117] "RemoveContainer" containerID="10dffd0c3394323fc7bcc1d00c465657f5ce32931dbdb98fa35ca149180eec75" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.951743 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b652247-e8c0-4986-9d98-156703326dee","Type":"ContainerDied","Data":"58995773b2c9f34f86487a366700fead2864f46948aea80075ea5849cdbf5ae0"} Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.951979 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerName="dnsmasq-dns" containerID="cri-o://92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717" gracePeriod=10 Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.952336 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.968372 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:33:53 crc kubenswrapper[4758]: I1203 18:33:53.983131 4758 scope.go:117] "RemoveContainer" containerID="bd02c354adfd86efb5b82889faf36bec8dac63cdf5fa6ddba2ef2b5c75c9a449" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.074348 4758 scope.go:117] "RemoveContainer" containerID="544eb92443af687a456ca0efc9fa52b4ad48c28f44e325f466f149b0fea393b2" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.105052 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.118746 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.129466 4758 scope.go:117] "RemoveContainer" containerID="4beb4de623e45e953a79c9bf05163519762a9ce4a1d730f676284f6a1b5d3519" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.149144 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: E1203 18:33:54.149847 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-metadata" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.149873 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-metadata" Dec 03 18:33:54 crc kubenswrapper[4758]: E1203 18:33:54.149897 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87223ea-1a03-43c6-bbe0-e9a11b4db504" containerName="nova-manage" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.149907 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87223ea-1a03-43c6-bbe0-e9a11b4db504" containerName="nova-manage" Dec 03 18:33:54 crc kubenswrapper[4758]: E1203 18:33:54.149926 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-log" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.149932 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-log" Dec 03 18:33:54 crc kubenswrapper[4758]: E1203 18:33:54.149943 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-log" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.149950 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-log" Dec 03 18:33:54 crc kubenswrapper[4758]: E1203 18:33:54.149993 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-api" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.150002 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-api" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.150228 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-metadata" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.150257 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87223ea-1a03-43c6-bbe0-e9a11b4db504" containerName="nova-manage" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.150269 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-api" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.150290 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" containerName="nova-api-log" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.150306 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b652247-e8c0-4986-9d98-156703326dee" containerName="nova-metadata-log" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.155156 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.158486 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.183384 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.191505 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.202155 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.211612 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.213507 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.218789 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.220758 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.238141 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq6pg\" (UniqueName: \"kubernetes.io/projected/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-kube-api-access-qq6pg\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.238248 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-logs\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.238304 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.238354 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-config-data\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.344983 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223b807a-92c0-49ac-b76c-eef81b199d3b-logs\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.345280 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-config-data\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.346221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h5gq\" (UniqueName: \"kubernetes.io/projected/223b807a-92c0-49ac-b76c-eef81b199d3b-kube-api-access-9h5gq\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.346258 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-config-data\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.346669 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq6pg\" (UniqueName: \"kubernetes.io/projected/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-kube-api-access-qq6pg\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.346764 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-logs\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.346823 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.346847 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.347498 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-logs\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.352061 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-config-data\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.365515 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.371656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq6pg\" (UniqueName: \"kubernetes.io/projected/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-kube-api-access-qq6pg\") pod \"nova-metadata-0\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.448742 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.448799 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223b807a-92c0-49ac-b76c-eef81b199d3b-logs\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.448851 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h5gq\" (UniqueName: \"kubernetes.io/projected/223b807a-92c0-49ac-b76c-eef81b199d3b-kube-api-access-9h5gq\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.448872 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-config-data\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.449537 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223b807a-92c0-49ac-b76c-eef81b199d3b-logs\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.455416 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-config-data\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.459598 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.464115 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h5gq\" (UniqueName: \"kubernetes.io/projected/223b807a-92c0-49ac-b76c-eef81b199d3b-kube-api-access-9h5gq\") pod \"nova-api-0\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.520295 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.531085 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.556661 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.652497 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-config\") pod \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.652571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-nb\") pod \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.652598 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-sb\") pod \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.652673 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-dns-svc\") pod \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.652775 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk6p4\" (UniqueName: \"kubernetes.io/projected/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-kube-api-access-mk6p4\") pod \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\" (UID: \"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2\") " Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.656348 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-kube-api-access-mk6p4" (OuterVolumeSpecName: "kube-api-access-mk6p4") pod "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" (UID: "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2"). InnerVolumeSpecName "kube-api-access-mk6p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.696826 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-config" (OuterVolumeSpecName: "config") pod "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" (UID: "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.698338 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" (UID: "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.701542 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" (UID: "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.710016 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" (UID: "cde89c0e-afcd-4b4f-87b8-a3bb61d904e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.754827 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.754895 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.754909 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.754921 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.754933 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk6p4\" (UniqueName: \"kubernetes.io/projected/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2-kube-api-access-mk6p4\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.973061 4758 generic.go:334] "Generic (PLEG): container finished" podID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerID="92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717" exitCode=0 Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.973423 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.973277 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" event={"ID":"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2","Type":"ContainerDied","Data":"92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717"} Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.975338 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5584c87ddf-q6bzc" event={"ID":"cde89c0e-afcd-4b4f-87b8-a3bb61d904e2","Type":"ContainerDied","Data":"e66b5f4cafbed5224144232d6c2892b23162b94ad643fc03e275b83d14a529d7"} Dec 03 18:33:54 crc kubenswrapper[4758]: I1203 18:33:54.975366 4758 scope.go:117] "RemoveContainer" containerID="92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.016752 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.045752 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.048813 4758 scope.go:117] "RemoveContainer" containerID="9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.069609 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5584c87ddf-q6bzc"] Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.082179 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5584c87ddf-q6bzc"] Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.115835 4758 scope.go:117] "RemoveContainer" containerID="92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717" Dec 03 18:33:55 crc kubenswrapper[4758]: E1203 18:33:55.128836 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717\": container with ID starting with 92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717 not found: ID does not exist" containerID="92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.128892 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717"} err="failed to get container status \"92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717\": rpc error: code = NotFound desc = could not find container \"92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717\": container with ID starting with 92c4a75bbef4e7f90373f289bedc81b9f7d30c377220789e07f750b7aa49d717 not found: ID does not exist" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.128921 4758 scope.go:117] "RemoveContainer" containerID="9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9" Dec 03 18:33:55 crc kubenswrapper[4758]: E1203 18:33:55.132897 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9\": container with ID starting with 9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9 not found: ID does not exist" containerID="9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.132950 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9"} err="failed to get container status \"9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9\": rpc error: code = NotFound desc = could not find container \"9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9\": container with ID starting with 9a5109b47fc17b496cf5c939de7a13327c4714c2acf4d852d7ac593bf593c1d9 not found: ID does not exist" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.138532 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b652247-e8c0-4986-9d98-156703326dee" path="/var/lib/kubelet/pods/3b652247-e8c0-4986-9d98-156703326dee/volumes" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.139508 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfa37d80-1e92-4a29-82f2-66f902c2021c" path="/var/lib/kubelet/pods/bfa37d80-1e92-4a29-82f2-66f902c2021c/volumes" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.142855 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" path="/var/lib/kubelet/pods/cde89c0e-afcd-4b4f-87b8-a3bb61d904e2/volumes" Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.986542 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"223b807a-92c0-49ac-b76c-eef81b199d3b","Type":"ContainerStarted","Data":"eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e"} Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.987951 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"223b807a-92c0-49ac-b76c-eef81b199d3b","Type":"ContainerStarted","Data":"773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14"} Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.988072 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"223b807a-92c0-49ac-b76c-eef81b199d3b","Type":"ContainerStarted","Data":"ccb19dca0b1f7d16e166fa580a71d663f42b2bf47620f5aed27bacc4d6f92c08"} Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.989917 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd0b0b3f-4989-4d00-b574-6751d8b43f8f","Type":"ContainerStarted","Data":"be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030"} Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.990210 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd0b0b3f-4989-4d00-b574-6751d8b43f8f","Type":"ContainerStarted","Data":"1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3"} Dec 03 18:33:55 crc kubenswrapper[4758]: I1203 18:33:55.990231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd0b0b3f-4989-4d00-b574-6751d8b43f8f","Type":"ContainerStarted","Data":"c6c183d369a807197d0ed1db9977a67ea85c145bcca4eedf890e0eca6304d1b1"} Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.011465 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.011445842 podStartE2EDuration="2.011445842s" podCreationTimestamp="2025-12-03 18:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:56.005126462 +0000 UTC m=+5891.206503383" watchObservedRunningTime="2025-12-03 18:33:56.011445842 +0000 UTC m=+5891.212822693" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.033811 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.033789782 podStartE2EDuration="2.033789782s" podCreationTimestamp="2025-12-03 18:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:56.031915622 +0000 UTC m=+5891.233292513" watchObservedRunningTime="2025-12-03 18:33:56.033789782 +0000 UTC m=+5891.235166643" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.424844 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.600270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgqjv\" (UniqueName: \"kubernetes.io/projected/2828e8f9-352b-4718-990e-e9db59f6ba4a-kube-api-access-mgqjv\") pod \"2828e8f9-352b-4718-990e-e9db59f6ba4a\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.600539 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-combined-ca-bundle\") pod \"2828e8f9-352b-4718-990e-e9db59f6ba4a\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.600576 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-config-data\") pod \"2828e8f9-352b-4718-990e-e9db59f6ba4a\" (UID: \"2828e8f9-352b-4718-990e-e9db59f6ba4a\") " Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.607064 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2828e8f9-352b-4718-990e-e9db59f6ba4a-kube-api-access-mgqjv" (OuterVolumeSpecName: "kube-api-access-mgqjv") pod "2828e8f9-352b-4718-990e-e9db59f6ba4a" (UID: "2828e8f9-352b-4718-990e-e9db59f6ba4a"). InnerVolumeSpecName "kube-api-access-mgqjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.628154 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2828e8f9-352b-4718-990e-e9db59f6ba4a" (UID: "2828e8f9-352b-4718-990e-e9db59f6ba4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.655733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-config-data" (OuterVolumeSpecName: "config-data") pod "2828e8f9-352b-4718-990e-e9db59f6ba4a" (UID: "2828e8f9-352b-4718-990e-e9db59f6ba4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.704163 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgqjv\" (UniqueName: \"kubernetes.io/projected/2828e8f9-352b-4718-990e-e9db59f6ba4a-kube-api-access-mgqjv\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.704236 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:56 crc kubenswrapper[4758]: I1203 18:33:56.704258 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2828e8f9-352b-4718-990e-e9db59f6ba4a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.004923 4758 generic.go:334] "Generic (PLEG): container finished" podID="2828e8f9-352b-4718-990e-e9db59f6ba4a" containerID="bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f" exitCode=0 Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.004984 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2828e8f9-352b-4718-990e-e9db59f6ba4a","Type":"ContainerDied","Data":"bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f"} Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.005046 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.005117 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2828e8f9-352b-4718-990e-e9db59f6ba4a","Type":"ContainerDied","Data":"ca032caaf97420fcdb37518e602b1baa1e546a683baffa70667096665d842291"} Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.005165 4758 scope.go:117] "RemoveContainer" containerID="bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.040357 4758 scope.go:117] "RemoveContainer" containerID="bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f" Dec 03 18:33:57 crc kubenswrapper[4758]: E1203 18:33:57.043397 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f\": container with ID starting with bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f not found: ID does not exist" containerID="bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.043456 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f"} err="failed to get container status \"bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f\": rpc error: code = NotFound desc = could not find container \"bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f\": container with ID starting with bd7a6975d237714bbe079c5282d7a202b9953e261433978b62c3c0155df9039f not found: ID does not exist" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.053905 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.071942 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.087651 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:57 crc kubenswrapper[4758]: E1203 18:33:57.088132 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerName="dnsmasq-dns" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.088153 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerName="dnsmasq-dns" Dec 03 18:33:57 crc kubenswrapper[4758]: E1203 18:33:57.088180 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2828e8f9-352b-4718-990e-e9db59f6ba4a" containerName="nova-scheduler-scheduler" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.088188 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2828e8f9-352b-4718-990e-e9db59f6ba4a" containerName="nova-scheduler-scheduler" Dec 03 18:33:57 crc kubenswrapper[4758]: E1203 18:33:57.088200 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerName="init" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.088210 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerName="init" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.088466 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2828e8f9-352b-4718-990e-e9db59f6ba4a" containerName="nova-scheduler-scheduler" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.088486 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde89c0e-afcd-4b4f-87b8-a3bb61d904e2" containerName="dnsmasq-dns" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.089248 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.096388 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.097377 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.113239 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-config-data\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.113637 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.113748 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqmtg\" (UniqueName: \"kubernetes.io/projected/f86390e0-2bd7-428e-8f93-590fac302849-kube-api-access-fqmtg\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.126799 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2828e8f9-352b-4718-990e-e9db59f6ba4a" path="/var/lib/kubelet/pods/2828e8f9-352b-4718-990e-e9db59f6ba4a/volumes" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.215470 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-config-data\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.215567 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.215613 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqmtg\" (UniqueName: \"kubernetes.io/projected/f86390e0-2bd7-428e-8f93-590fac302849-kube-api-access-fqmtg\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.221309 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.230130 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-config-data\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.232415 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqmtg\" (UniqueName: \"kubernetes.io/projected/f86390e0-2bd7-428e-8f93-590fac302849-kube-api-access-fqmtg\") pod \"nova-scheduler-0\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.409320 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:33:57 crc kubenswrapper[4758]: I1203 18:33:57.858988 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:33:58 crc kubenswrapper[4758]: I1203 18:33:58.014493 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f86390e0-2bd7-428e-8f93-590fac302849","Type":"ContainerStarted","Data":"85534db83444f1d5b87c6cd7d707e84af890489ae3aa00e458a3593088ab0d72"} Dec 03 18:33:59 crc kubenswrapper[4758]: I1203 18:33:59.026152 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f86390e0-2bd7-428e-8f93-590fac302849","Type":"ContainerStarted","Data":"879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575"} Dec 03 18:33:59 crc kubenswrapper[4758]: I1203 18:33:59.046986 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.046966785 podStartE2EDuration="2.046966785s" podCreationTimestamp="2025-12-03 18:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:33:59.041133879 +0000 UTC m=+5894.242510750" watchObservedRunningTime="2025-12-03 18:33:59.046966785 +0000 UTC m=+5894.248343646" Dec 03 18:33:59 crc kubenswrapper[4758]: I1203 18:33:59.531443 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:33:59 crc kubenswrapper[4758]: I1203 18:33:59.531512 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.114458 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:34:01 crc kubenswrapper[4758]: E1203 18:34:01.115020 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.359124 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.878257 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-b55f9"] Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.879580 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.882214 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.882616 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.889717 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-b55f9"] Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.901021 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvbvb\" (UniqueName: \"kubernetes.io/projected/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-kube-api-access-qvbvb\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.901136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.901176 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-scripts\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:01 crc kubenswrapper[4758]: I1203 18:34:01.901206 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-config-data\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.002832 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvbvb\" (UniqueName: \"kubernetes.io/projected/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-kube-api-access-qvbvb\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.002918 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.002948 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-scripts\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.002972 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-config-data\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.009564 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-config-data\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.010381 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.010449 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-scripts\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.022128 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvbvb\" (UniqueName: \"kubernetes.io/projected/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-kube-api-access-qvbvb\") pod \"nova-cell1-cell-mapping-b55f9\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.262972 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.410425 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 18:34:02 crc kubenswrapper[4758]: W1203 18:34:02.712747 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c7eeb6e_1676_45f8_9f6d_c704c0be486f.slice/crio-62616de02ce103009fbf62cfc691e9772a92be5550d1b8685efcdf74e796b76c WatchSource:0}: Error finding container 62616de02ce103009fbf62cfc691e9772a92be5550d1b8685efcdf74e796b76c: Status 404 returned error can't find the container with id 62616de02ce103009fbf62cfc691e9772a92be5550d1b8685efcdf74e796b76c Dec 03 18:34:02 crc kubenswrapper[4758]: I1203 18:34:02.722397 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-b55f9"] Dec 03 18:34:03 crc kubenswrapper[4758]: I1203 18:34:03.066099 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b55f9" event={"ID":"8c7eeb6e-1676-45f8-9f6d-c704c0be486f","Type":"ContainerStarted","Data":"62616de02ce103009fbf62cfc691e9772a92be5550d1b8685efcdf74e796b76c"} Dec 03 18:34:04 crc kubenswrapper[4758]: I1203 18:34:04.085197 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b55f9" event={"ID":"8c7eeb6e-1676-45f8-9f6d-c704c0be486f","Type":"ContainerStarted","Data":"2946b52948591e4ccd52bdd8de2e84e01e9e09d32b1a02f966c1306b5bc08bf4"} Dec 03 18:34:04 crc kubenswrapper[4758]: I1203 18:34:04.107180 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-b55f9" podStartSLOduration=3.107136701 podStartE2EDuration="3.107136701s" podCreationTimestamp="2025-12-03 18:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:34:04.10373334 +0000 UTC m=+5899.305110221" watchObservedRunningTime="2025-12-03 18:34:04.107136701 +0000 UTC m=+5899.308513562" Dec 03 18:34:04 crc kubenswrapper[4758]: I1203 18:34:04.531349 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 18:34:04 crc kubenswrapper[4758]: I1203 18:34:04.531655 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 18:34:04 crc kubenswrapper[4758]: I1203 18:34:04.557576 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 18:34:04 crc kubenswrapper[4758]: I1203 18:34:04.558529 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 18:34:05 crc kubenswrapper[4758]: I1203 18:34:05.697922 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:05 crc kubenswrapper[4758]: I1203 18:34:05.697959 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:05 crc kubenswrapper[4758]: I1203 18:34:05.697930 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:05 crc kubenswrapper[4758]: I1203 18:34:05.697932 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:07 crc kubenswrapper[4758]: I1203 18:34:07.410406 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 18:34:07 crc kubenswrapper[4758]: I1203 18:34:07.436422 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 18:34:08 crc kubenswrapper[4758]: I1203 18:34:08.143105 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 18:34:09 crc kubenswrapper[4758]: I1203 18:34:09.140883 4758 generic.go:334] "Generic (PLEG): container finished" podID="8c7eeb6e-1676-45f8-9f6d-c704c0be486f" containerID="2946b52948591e4ccd52bdd8de2e84e01e9e09d32b1a02f966c1306b5bc08bf4" exitCode=0 Dec 03 18:34:09 crc kubenswrapper[4758]: I1203 18:34:09.141113 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b55f9" event={"ID":"8c7eeb6e-1676-45f8-9f6d-c704c0be486f","Type":"ContainerDied","Data":"2946b52948591e4ccd52bdd8de2e84e01e9e09d32b1a02f966c1306b5bc08bf4"} Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.500138 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.689707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-combined-ca-bundle\") pod \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.689890 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvbvb\" (UniqueName: \"kubernetes.io/projected/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-kube-api-access-qvbvb\") pod \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.689960 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-config-data\") pod \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.689979 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-scripts\") pod \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\" (UID: \"8c7eeb6e-1676-45f8-9f6d-c704c0be486f\") " Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.695346 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-scripts" (OuterVolumeSpecName: "scripts") pod "8c7eeb6e-1676-45f8-9f6d-c704c0be486f" (UID: "8c7eeb6e-1676-45f8-9f6d-c704c0be486f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.695405 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-kube-api-access-qvbvb" (OuterVolumeSpecName: "kube-api-access-qvbvb") pod "8c7eeb6e-1676-45f8-9f6d-c704c0be486f" (UID: "8c7eeb6e-1676-45f8-9f6d-c704c0be486f"). InnerVolumeSpecName "kube-api-access-qvbvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.714716 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-config-data" (OuterVolumeSpecName: "config-data") pod "8c7eeb6e-1676-45f8-9f6d-c704c0be486f" (UID: "8c7eeb6e-1676-45f8-9f6d-c704c0be486f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.724117 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c7eeb6e-1676-45f8-9f6d-c704c0be486f" (UID: "8c7eeb6e-1676-45f8-9f6d-c704c0be486f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.791737 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvbvb\" (UniqueName: \"kubernetes.io/projected/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-kube-api-access-qvbvb\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.791774 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.791785 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:10 crc kubenswrapper[4758]: I1203 18:34:10.791794 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7eeb6e-1676-45f8-9f6d-c704c0be486f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.159736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b55f9" event={"ID":"8c7eeb6e-1676-45f8-9f6d-c704c0be486f","Type":"ContainerDied","Data":"62616de02ce103009fbf62cfc691e9772a92be5550d1b8685efcdf74e796b76c"} Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.160096 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62616de02ce103009fbf62cfc691e9772a92be5550d1b8685efcdf74e796b76c" Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.159799 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b55f9" Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.339718 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.339977 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-log" containerID="cri-o://773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14" gracePeriod=30 Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.340124 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-api" containerID="cri-o://eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e" gracePeriod=30 Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.363109 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.363288 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f86390e0-2bd7-428e-8f93-590fac302849" containerName="nova-scheduler-scheduler" containerID="cri-o://879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" gracePeriod=30 Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.389730 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.390013 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-log" containerID="cri-o://1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3" gracePeriod=30 Dec 03 18:34:11 crc kubenswrapper[4758]: I1203 18:34:11.390417 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-metadata" containerID="cri-o://be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030" gracePeriod=30 Dec 03 18:34:12 crc kubenswrapper[4758]: I1203 18:34:12.169191 4758 generic.go:334] "Generic (PLEG): container finished" podID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerID="773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14" exitCode=143 Dec 03 18:34:12 crc kubenswrapper[4758]: I1203 18:34:12.169269 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"223b807a-92c0-49ac-b76c-eef81b199d3b","Type":"ContainerDied","Data":"773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14"} Dec 03 18:34:12 crc kubenswrapper[4758]: I1203 18:34:12.171245 4758 generic.go:334] "Generic (PLEG): container finished" podID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerID="1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3" exitCode=143 Dec 03 18:34:12 crc kubenswrapper[4758]: I1203 18:34:12.171276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd0b0b3f-4989-4d00-b574-6751d8b43f8f","Type":"ContainerDied","Data":"1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3"} Dec 03 18:34:12 crc kubenswrapper[4758]: E1203 18:34:12.411091 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 18:34:12 crc kubenswrapper[4758]: E1203 18:34:12.413424 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 18:34:12 crc kubenswrapper[4758]: E1203 18:34:12.415091 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 18:34:12 crc kubenswrapper[4758]: E1203 18:34:12.415159 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f86390e0-2bd7-428e-8f93-590fac302849" containerName="nova-scheduler-scheduler" Dec 03 18:34:12 crc kubenswrapper[4758]: I1203 18:34:12.624312 4758 scope.go:117] "RemoveContainer" containerID="9d51de49bcf89ec563eabbe282e0f7116aaeae6a4093531e584b68bdc46b22a9" Dec 03 18:34:13 crc kubenswrapper[4758]: I1203 18:34:13.114375 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:34:13 crc kubenswrapper[4758]: E1203 18:34:13.114863 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.743952 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.865612 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.872517 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqmtg\" (UniqueName: \"kubernetes.io/projected/f86390e0-2bd7-428e-8f93-590fac302849-kube-api-access-fqmtg\") pod \"f86390e0-2bd7-428e-8f93-590fac302849\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.872752 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-combined-ca-bundle\") pod \"f86390e0-2bd7-428e-8f93-590fac302849\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.872885 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-config-data\") pod \"f86390e0-2bd7-428e-8f93-590fac302849\" (UID: \"f86390e0-2bd7-428e-8f93-590fac302849\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.878281 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f86390e0-2bd7-428e-8f93-590fac302849-kube-api-access-fqmtg" (OuterVolumeSpecName: "kube-api-access-fqmtg") pod "f86390e0-2bd7-428e-8f93-590fac302849" (UID: "f86390e0-2bd7-428e-8f93-590fac302849"). InnerVolumeSpecName "kube-api-access-fqmtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.924793 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.932319 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-config-data" (OuterVolumeSpecName: "config-data") pod "f86390e0-2bd7-428e-8f93-590fac302849" (UID: "f86390e0-2bd7-428e-8f93-590fac302849"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.934823 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f86390e0-2bd7-428e-8f93-590fac302849" (UID: "f86390e0-2bd7-428e-8f93-590fac302849"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975040 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-combined-ca-bundle\") pod \"223b807a-92c0-49ac-b76c-eef81b199d3b\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975195 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h5gq\" (UniqueName: \"kubernetes.io/projected/223b807a-92c0-49ac-b76c-eef81b199d3b-kube-api-access-9h5gq\") pod \"223b807a-92c0-49ac-b76c-eef81b199d3b\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975325 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-config-data\") pod \"223b807a-92c0-49ac-b76c-eef81b199d3b\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975373 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223b807a-92c0-49ac-b76c-eef81b199d3b-logs\") pod \"223b807a-92c0-49ac-b76c-eef81b199d3b\" (UID: \"223b807a-92c0-49ac-b76c-eef81b199d3b\") " Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975707 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqmtg\" (UniqueName: \"kubernetes.io/projected/f86390e0-2bd7-428e-8f93-590fac302849-kube-api-access-fqmtg\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975729 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975741 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86390e0-2bd7-428e-8f93-590fac302849-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.975998 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/223b807a-92c0-49ac-b76c-eef81b199d3b-logs" (OuterVolumeSpecName: "logs") pod "223b807a-92c0-49ac-b76c-eef81b199d3b" (UID: "223b807a-92c0-49ac-b76c-eef81b199d3b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.978320 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223b807a-92c0-49ac-b76c-eef81b199d3b-kube-api-access-9h5gq" (OuterVolumeSpecName: "kube-api-access-9h5gq") pod "223b807a-92c0-49ac-b76c-eef81b199d3b" (UID: "223b807a-92c0-49ac-b76c-eef81b199d3b"). InnerVolumeSpecName "kube-api-access-9h5gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.996109 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-config-data" (OuterVolumeSpecName: "config-data") pod "223b807a-92c0-49ac-b76c-eef81b199d3b" (UID: "223b807a-92c0-49ac-b76c-eef81b199d3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:14 crc kubenswrapper[4758]: I1203 18:34:14.997008 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "223b807a-92c0-49ac-b76c-eef81b199d3b" (UID: "223b807a-92c0-49ac-b76c-eef81b199d3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.076938 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-combined-ca-bundle\") pod \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.077012 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-logs\") pod \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.077136 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq6pg\" (UniqueName: \"kubernetes.io/projected/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-kube-api-access-qq6pg\") pod \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.077311 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-config-data\") pod \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\" (UID: \"dd0b0b3f-4989-4d00-b574-6751d8b43f8f\") " Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.077555 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-logs" (OuterVolumeSpecName: "logs") pod "dd0b0b3f-4989-4d00-b574-6751d8b43f8f" (UID: "dd0b0b3f-4989-4d00-b574-6751d8b43f8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.078071 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223b807a-92c0-49ac-b76c-eef81b199d3b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.078097 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.078109 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h5gq\" (UniqueName: \"kubernetes.io/projected/223b807a-92c0-49ac-b76c-eef81b199d3b-kube-api-access-9h5gq\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.078118 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.078130 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223b807a-92c0-49ac-b76c-eef81b199d3b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.080026 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-kube-api-access-qq6pg" (OuterVolumeSpecName: "kube-api-access-qq6pg") pod "dd0b0b3f-4989-4d00-b574-6751d8b43f8f" (UID: "dd0b0b3f-4989-4d00-b574-6751d8b43f8f"). InnerVolumeSpecName "kube-api-access-qq6pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.098482 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-config-data" (OuterVolumeSpecName: "config-data") pod "dd0b0b3f-4989-4d00-b574-6751d8b43f8f" (UID: "dd0b0b3f-4989-4d00-b574-6751d8b43f8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.100244 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd0b0b3f-4989-4d00-b574-6751d8b43f8f" (UID: "dd0b0b3f-4989-4d00-b574-6751d8b43f8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.180064 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.180250 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.180296 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq6pg\" (UniqueName: \"kubernetes.io/projected/dd0b0b3f-4989-4d00-b574-6751d8b43f8f-kube-api-access-qq6pg\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.199447 4758 generic.go:334] "Generic (PLEG): container finished" podID="f86390e0-2bd7-428e-8f93-590fac302849" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" exitCode=0 Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.199525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f86390e0-2bd7-428e-8f93-590fac302849","Type":"ContainerDied","Data":"879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575"} Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.199537 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.199555 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f86390e0-2bd7-428e-8f93-590fac302849","Type":"ContainerDied","Data":"85534db83444f1d5b87c6cd7d707e84af890489ae3aa00e458a3593088ab0d72"} Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.199580 4758 scope.go:117] "RemoveContainer" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.201045 4758 generic.go:334] "Generic (PLEG): container finished" podID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerID="eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e" exitCode=0 Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.201083 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"223b807a-92c0-49ac-b76c-eef81b199d3b","Type":"ContainerDied","Data":"eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e"} Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.201099 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"223b807a-92c0-49ac-b76c-eef81b199d3b","Type":"ContainerDied","Data":"ccb19dca0b1f7d16e166fa580a71d663f42b2bf47620f5aed27bacc4d6f92c08"} Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.201135 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.205575 4758 generic.go:334] "Generic (PLEG): container finished" podID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerID="be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030" exitCode=0 Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.205621 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.205648 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd0b0b3f-4989-4d00-b574-6751d8b43f8f","Type":"ContainerDied","Data":"be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030"} Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.206099 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd0b0b3f-4989-4d00-b574-6751d8b43f8f","Type":"ContainerDied","Data":"c6c183d369a807197d0ed1db9977a67ea85c145bcca4eedf890e0eca6304d1b1"} Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.225438 4758 scope.go:117] "RemoveContainer" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.226510 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575\": container with ID starting with 879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575 not found: ID does not exist" containerID="879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.226548 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575"} err="failed to get container status \"879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575\": rpc error: code = NotFound desc = could not find container \"879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575\": container with ID starting with 879f4697a7ab2437daa1789345aebe443b81ee72a39cf8d7e56ec9b6ce9d8575 not found: ID does not exist" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.226570 4758 scope.go:117] "RemoveContainer" containerID="eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.245790 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.267061 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.279403 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.280585 4758 scope.go:117] "RemoveContainer" containerID="773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.308529 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328137 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.328738 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-metadata" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328759 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-metadata" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.328786 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-api" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328793 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-api" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.328814 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-log" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328820 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-log" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.328830 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-log" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328837 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-log" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.328854 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f86390e0-2bd7-428e-8f93-590fac302849" containerName="nova-scheduler-scheduler" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328860 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f86390e0-2bd7-428e-8f93-590fac302849" containerName="nova-scheduler-scheduler" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.328867 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7eeb6e-1676-45f8-9f6d-c704c0be486f" containerName="nova-manage" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.328872 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7eeb6e-1676-45f8-9f6d-c704c0be486f" containerName="nova-manage" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329039 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-log" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329059 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-metadata" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329074 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" containerName="nova-api-api" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329081 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7eeb6e-1676-45f8-9f6d-c704c0be486f" containerName="nova-manage" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329089 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" containerName="nova-metadata-log" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329100 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f86390e0-2bd7-428e-8f93-590fac302849" containerName="nova-scheduler-scheduler" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.329748 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.332189 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.340560 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.350145 4758 scope.go:117] "RemoveContainer" containerID="eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.351147 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e\": container with ID starting with eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e not found: ID does not exist" containerID="eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.351238 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e"} err="failed to get container status \"eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e\": rpc error: code = NotFound desc = could not find container \"eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e\": container with ID starting with eb3bd5458a5ababceac941acfdcbc118f152cb30c67faffc64f2b0742a19f42e not found: ID does not exist" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.351346 4758 scope.go:117] "RemoveContainer" containerID="773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.351713 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14\": container with ID starting with 773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14 not found: ID does not exist" containerID="773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.351751 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14"} err="failed to get container status \"773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14\": rpc error: code = NotFound desc = could not find container \"773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14\": container with ID starting with 773712805a60c40ed6019ddbb324611bc8c2aca862033e62c973f7499d8b4c14 not found: ID does not exist" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.351778 4758 scope.go:117] "RemoveContainer" containerID="be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.359396 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.367575 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.369452 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.371878 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.376178 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.378586 4758 scope.go:117] "RemoveContainer" containerID="1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.388491 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.397641 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.399182 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.401312 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.403591 4758 scope.go:117] "RemoveContainer" containerID="be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.404146 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030\": container with ID starting with be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030 not found: ID does not exist" containerID="be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.404195 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030"} err="failed to get container status \"be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030\": rpc error: code = NotFound desc = could not find container \"be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030\": container with ID starting with be6a89c9065146f6857ea637cf1e45a08c2e2bacd7ebc48290ad3dba33aeb030 not found: ID does not exist" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.404225 4758 scope.go:117] "RemoveContainer" containerID="1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3" Dec 03 18:34:15 crc kubenswrapper[4758]: E1203 18:34:15.404610 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3\": container with ID starting with 1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3 not found: ID does not exist" containerID="1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.404640 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3"} err="failed to get container status \"1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3\": rpc error: code = NotFound desc = could not find container \"1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3\": container with ID starting with 1ea09c1a7f749ff6aff0c5b892bf293c0b054ab5a43548aa4c83280f99837ce3 not found: ID does not exist" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.405344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.488714 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a192022-6125-4f06-aefc-b65c68c931d0-logs\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.488766 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.489698 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.489750 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htl54\" (UniqueName: \"kubernetes.io/projected/941980fe-3b3a-4183-bbbe-08a7d71563a6-kube-api-access-htl54\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.489845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2r2b\" (UniqueName: \"kubernetes.io/projected/3a192022-6125-4f06-aefc-b65c68c931d0-kube-api-access-z2r2b\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.489993 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-config-data\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.490278 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-config-data\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592199 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-config-data\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592313 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-config-data\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592365 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a192022-6125-4f06-aefc-b65c68c931d0-logs\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592452 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-logs\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592525 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592572 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htl54\" (UniqueName: \"kubernetes.io/projected/941980fe-3b3a-4183-bbbe-08a7d71563a6-kube-api-access-htl54\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592658 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592698 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkxfn\" (UniqueName: \"kubernetes.io/projected/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-kube-api-access-vkxfn\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592758 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2r2b\" (UniqueName: \"kubernetes.io/projected/3a192022-6125-4f06-aefc-b65c68c931d0-kube-api-access-z2r2b\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.592804 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-config-data\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.593982 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a192022-6125-4f06-aefc-b65c68c931d0-logs\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.598300 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-config-data\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.603637 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.604745 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-config-data\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.608091 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.612831 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htl54\" (UniqueName: \"kubernetes.io/projected/941980fe-3b3a-4183-bbbe-08a7d71563a6-kube-api-access-htl54\") pod \"nova-scheduler-0\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.613102 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2r2b\" (UniqueName: \"kubernetes.io/projected/3a192022-6125-4f06-aefc-b65c68c931d0-kube-api-access-z2r2b\") pod \"nova-api-0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.646328 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.693157 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.693820 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-config-data\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.693871 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-logs\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.693929 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.693946 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkxfn\" (UniqueName: \"kubernetes.io/projected/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-kube-api-access-vkxfn\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.697864 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-logs\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.698049 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-config-data\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.698527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.716439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkxfn\" (UniqueName: \"kubernetes.io/projected/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-kube-api-access-vkxfn\") pod \"nova-metadata-0\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " pod="openstack/nova-metadata-0" Dec 03 18:34:15 crc kubenswrapper[4758]: I1203 18:34:15.720395 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:34:16 crc kubenswrapper[4758]: I1203 18:34:16.091292 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:34:16 crc kubenswrapper[4758]: W1203 18:34:16.094622 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod941980fe_3b3a_4183_bbbe_08a7d71563a6.slice/crio-dd4fe19eb80298348193fc2d12e4ea2a172cf315ffda6a1778f5699b409db7a5 WatchSource:0}: Error finding container dd4fe19eb80298348193fc2d12e4ea2a172cf315ffda6a1778f5699b409db7a5: Status 404 returned error can't find the container with id dd4fe19eb80298348193fc2d12e4ea2a172cf315ffda6a1778f5699b409db7a5 Dec 03 18:34:16 crc kubenswrapper[4758]: I1203 18:34:16.175394 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:34:16 crc kubenswrapper[4758]: W1203 18:34:16.179778 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2d83593_6b04_4aeb_95b9_a0aa0de618eb.slice/crio-1cabb2cb6812518bd8419976cafa5ff5e3bbd9b68b0b9a7dc74a83bb1df22139 WatchSource:0}: Error finding container 1cabb2cb6812518bd8419976cafa5ff5e3bbd9b68b0b9a7dc74a83bb1df22139: Status 404 returned error can't find the container with id 1cabb2cb6812518bd8419976cafa5ff5e3bbd9b68b0b9a7dc74a83bb1df22139 Dec 03 18:34:16 crc kubenswrapper[4758]: W1203 18:34:16.181125 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a192022_6125_4f06_aefc_b65c68c931d0.slice/crio-e67cd50eb922d0c70a442eaeabcb4d760d9283317975d19af60c65a6e52991ea WatchSource:0}: Error finding container e67cd50eb922d0c70a442eaeabcb4d760d9283317975d19af60c65a6e52991ea: Status 404 returned error can't find the container with id e67cd50eb922d0c70a442eaeabcb4d760d9283317975d19af60c65a6e52991ea Dec 03 18:34:16 crc kubenswrapper[4758]: I1203 18:34:16.183863 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:34:16 crc kubenswrapper[4758]: I1203 18:34:16.218350 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"941980fe-3b3a-4183-bbbe-08a7d71563a6","Type":"ContainerStarted","Data":"dd4fe19eb80298348193fc2d12e4ea2a172cf315ffda6a1778f5699b409db7a5"} Dec 03 18:34:16 crc kubenswrapper[4758]: I1203 18:34:16.220294 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2d83593-6b04-4aeb-95b9-a0aa0de618eb","Type":"ContainerStarted","Data":"1cabb2cb6812518bd8419976cafa5ff5e3bbd9b68b0b9a7dc74a83bb1df22139"} Dec 03 18:34:16 crc kubenswrapper[4758]: I1203 18:34:16.221440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a192022-6125-4f06-aefc-b65c68c931d0","Type":"ContainerStarted","Data":"e67cd50eb922d0c70a442eaeabcb4d760d9283317975d19af60c65a6e52991ea"} Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.133386 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="223b807a-92c0-49ac-b76c-eef81b199d3b" path="/var/lib/kubelet/pods/223b807a-92c0-49ac-b76c-eef81b199d3b/volumes" Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.135660 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd0b0b3f-4989-4d00-b574-6751d8b43f8f" path="/var/lib/kubelet/pods/dd0b0b3f-4989-4d00-b574-6751d8b43f8f/volumes" Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.136847 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f86390e0-2bd7-428e-8f93-590fac302849" path="/var/lib/kubelet/pods/f86390e0-2bd7-428e-8f93-590fac302849/volumes" Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.236729 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a192022-6125-4f06-aefc-b65c68c931d0","Type":"ContainerStarted","Data":"ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9"} Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.236779 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a192022-6125-4f06-aefc-b65c68c931d0","Type":"ContainerStarted","Data":"22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28"} Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.240276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"941980fe-3b3a-4183-bbbe-08a7d71563a6","Type":"ContainerStarted","Data":"6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3"} Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.242991 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2d83593-6b04-4aeb-95b9-a0aa0de618eb","Type":"ContainerStarted","Data":"37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734"} Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.243037 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2d83593-6b04-4aeb-95b9-a0aa0de618eb","Type":"ContainerStarted","Data":"c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455"} Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.252358 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.252333081 podStartE2EDuration="2.252333081s" podCreationTimestamp="2025-12-03 18:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:34:17.249842714 +0000 UTC m=+5912.451219585" watchObservedRunningTime="2025-12-03 18:34:17.252333081 +0000 UTC m=+5912.453709952" Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.265976 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.265955078 podStartE2EDuration="2.265955078s" podCreationTimestamp="2025-12-03 18:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:34:17.262145215 +0000 UTC m=+5912.463522076" watchObservedRunningTime="2025-12-03 18:34:17.265955078 +0000 UTC m=+5912.467331939" Dec 03 18:34:17 crc kubenswrapper[4758]: I1203 18:34:17.292199 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.292179722 podStartE2EDuration="2.292179722s" podCreationTimestamp="2025-12-03 18:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:34:17.283646473 +0000 UTC m=+5912.485023334" watchObservedRunningTime="2025-12-03 18:34:17.292179722 +0000 UTC m=+5912.493556583" Dec 03 18:34:20 crc kubenswrapper[4758]: I1203 18:34:20.646597 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 18:34:20 crc kubenswrapper[4758]: I1203 18:34:20.721058 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:34:20 crc kubenswrapper[4758]: I1203 18:34:20.721382 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.122966 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:34:25 crc kubenswrapper[4758]: E1203 18:34:25.123968 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.647106 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.677044 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.694054 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.694096 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.721896 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 18:34:25 crc kubenswrapper[4758]: I1203 18:34:25.721987 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 18:34:26 crc kubenswrapper[4758]: I1203 18:34:26.340114 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 18:34:26 crc kubenswrapper[4758]: I1203 18:34:26.858897 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:26 crc kubenswrapper[4758]: I1203 18:34:26.858967 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.76:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:26 crc kubenswrapper[4758]: I1203 18:34:26.858991 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.76:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:26 crc kubenswrapper[4758]: I1203 18:34:26.858999 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.697600 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.698553 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.699575 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.703178 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.726824 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.728733 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 18:34:35 crc kubenswrapper[4758]: I1203 18:34:35.730203 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.416574 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.419581 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.420645 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.674418 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65fb5d464f-w9xqz"] Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.683990 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.698517 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65fb5d464f-w9xqz"] Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.802110 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-dns-svc\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.802486 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-nb\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.802640 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-sb\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.802785 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-config\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.802913 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ht2j\" (UniqueName: \"kubernetes.io/projected/21bc458f-4b56-4ed0-b3e4-fbad588d8669-kube-api-access-7ht2j\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.904721 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-dns-svc\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.905107 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-nb\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.905152 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-sb\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.905174 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-config\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.905194 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ht2j\" (UniqueName: \"kubernetes.io/projected/21bc458f-4b56-4ed0-b3e4-fbad588d8669-kube-api-access-7ht2j\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.905725 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-dns-svc\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.906232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-nb\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.906299 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-sb\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.906469 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-config\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:36 crc kubenswrapper[4758]: I1203 18:34:36.928488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ht2j\" (UniqueName: \"kubernetes.io/projected/21bc458f-4b56-4ed0-b3e4-fbad588d8669-kube-api-access-7ht2j\") pod \"dnsmasq-dns-65fb5d464f-w9xqz\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:37 crc kubenswrapper[4758]: I1203 18:34:37.008562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:37 crc kubenswrapper[4758]: I1203 18:34:37.491469 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65fb5d464f-w9xqz"] Dec 03 18:34:38 crc kubenswrapper[4758]: I1203 18:34:38.114235 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:34:38 crc kubenswrapper[4758]: E1203 18:34:38.114745 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:34:38 crc kubenswrapper[4758]: I1203 18:34:38.433452 4758 generic.go:334] "Generic (PLEG): container finished" podID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerID="bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654" exitCode=0 Dec 03 18:34:38 crc kubenswrapper[4758]: I1203 18:34:38.433498 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" event={"ID":"21bc458f-4b56-4ed0-b3e4-fbad588d8669","Type":"ContainerDied","Data":"bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654"} Dec 03 18:34:38 crc kubenswrapper[4758]: I1203 18:34:38.433983 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" event={"ID":"21bc458f-4b56-4ed0-b3e4-fbad588d8669","Type":"ContainerStarted","Data":"e6e8c03b04f027e0f4a77ad77f6e7814e915d116c7fc5a8271f6c1ee4a4ae071"} Dec 03 18:34:39 crc kubenswrapper[4758]: I1203 18:34:39.444660 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" event={"ID":"21bc458f-4b56-4ed0-b3e4-fbad588d8669","Type":"ContainerStarted","Data":"61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266"} Dec 03 18:34:39 crc kubenswrapper[4758]: I1203 18:34:39.445314 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:39 crc kubenswrapper[4758]: I1203 18:34:39.468891 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" podStartSLOduration=3.468872357 podStartE2EDuration="3.468872357s" podCreationTimestamp="2025-12-03 18:34:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:34:39.462464265 +0000 UTC m=+5934.663841126" watchObservedRunningTime="2025-12-03 18:34:39.468872357 +0000 UTC m=+5934.670249218" Dec 03 18:34:47 crc kubenswrapper[4758]: I1203 18:34:47.009883 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:34:47 crc kubenswrapper[4758]: I1203 18:34:47.098460 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7647579f87-n8tn8"] Dec 03 18:34:47 crc kubenswrapper[4758]: I1203 18:34:47.100215 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" podUID="21a13077-0749-4821-8735-06d720d6079c" containerName="dnsmasq-dns" containerID="cri-o://07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0" gracePeriod=10 Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.066989 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.221310 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-sb\") pod \"21a13077-0749-4821-8735-06d720d6079c\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.221388 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-dns-svc\") pod \"21a13077-0749-4821-8735-06d720d6079c\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.221425 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-nb\") pod \"21a13077-0749-4821-8735-06d720d6079c\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.221462 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-config\") pod \"21a13077-0749-4821-8735-06d720d6079c\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.221499 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g7p6\" (UniqueName: \"kubernetes.io/projected/21a13077-0749-4821-8735-06d720d6079c-kube-api-access-5g7p6\") pod \"21a13077-0749-4821-8735-06d720d6079c\" (UID: \"21a13077-0749-4821-8735-06d720d6079c\") " Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.234142 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a13077-0749-4821-8735-06d720d6079c-kube-api-access-5g7p6" (OuterVolumeSpecName: "kube-api-access-5g7p6") pod "21a13077-0749-4821-8735-06d720d6079c" (UID: "21a13077-0749-4821-8735-06d720d6079c"). InnerVolumeSpecName "kube-api-access-5g7p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.269151 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21a13077-0749-4821-8735-06d720d6079c" (UID: "21a13077-0749-4821-8735-06d720d6079c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.275424 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-config" (OuterVolumeSpecName: "config") pod "21a13077-0749-4821-8735-06d720d6079c" (UID: "21a13077-0749-4821-8735-06d720d6079c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.278410 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21a13077-0749-4821-8735-06d720d6079c" (UID: "21a13077-0749-4821-8735-06d720d6079c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.298323 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21a13077-0749-4821-8735-06d720d6079c" (UID: "21a13077-0749-4821-8735-06d720d6079c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.328030 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.328098 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.328114 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.328127 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g7p6\" (UniqueName: \"kubernetes.io/projected/21a13077-0749-4821-8735-06d720d6079c-kube-api-access-5g7p6\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.328140 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21a13077-0749-4821-8735-06d720d6079c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.526230 4758 generic.go:334] "Generic (PLEG): container finished" podID="21a13077-0749-4821-8735-06d720d6079c" containerID="07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0" exitCode=0 Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.526274 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" event={"ID":"21a13077-0749-4821-8735-06d720d6079c","Type":"ContainerDied","Data":"07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0"} Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.526305 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" event={"ID":"21a13077-0749-4821-8735-06d720d6079c","Type":"ContainerDied","Data":"d7c36ee6c34a2974a42f53cb3363165a2ec58fbc4a124da99930f00ff99d174d"} Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.526322 4758 scope.go:117] "RemoveContainer" containerID="07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.526491 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7647579f87-n8tn8" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.562989 4758 scope.go:117] "RemoveContainer" containerID="8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.565934 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7647579f87-n8tn8"] Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.574404 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7647579f87-n8tn8"] Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.579984 4758 scope.go:117] "RemoveContainer" containerID="07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0" Dec 03 18:34:48 crc kubenswrapper[4758]: E1203 18:34:48.580662 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0\": container with ID starting with 07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0 not found: ID does not exist" containerID="07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.580857 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0"} err="failed to get container status \"07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0\": rpc error: code = NotFound desc = could not find container \"07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0\": container with ID starting with 07684edd65e2ea1efc5fedbbb9a815beeb32a1909df64ef3e9e63d7976fa4cf0 not found: ID does not exist" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.580994 4758 scope.go:117] "RemoveContainer" containerID="8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0" Dec 03 18:34:48 crc kubenswrapper[4758]: E1203 18:34:48.581387 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0\": container with ID starting with 8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0 not found: ID does not exist" containerID="8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0" Dec 03 18:34:48 crc kubenswrapper[4758]: I1203 18:34:48.581511 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0"} err="failed to get container status \"8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0\": rpc error: code = NotFound desc = could not find container \"8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0\": container with ID starting with 8356f21f053699e7932c8a2dd612801263afce64038d24b06cc4a094a924e5d0 not found: ID does not exist" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.125548 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a13077-0749-4821-8735-06d720d6079c" path="/var/lib/kubelet/pods/21a13077-0749-4821-8735-06d720d6079c/volumes" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.723411 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rtkk7"] Dec 03 18:34:49 crc kubenswrapper[4758]: E1203 18:34:49.723858 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a13077-0749-4821-8735-06d720d6079c" containerName="dnsmasq-dns" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.723877 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a13077-0749-4821-8735-06d720d6079c" containerName="dnsmasq-dns" Dec 03 18:34:49 crc kubenswrapper[4758]: E1203 18:34:49.723895 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a13077-0749-4821-8735-06d720d6079c" containerName="init" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.723903 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a13077-0749-4821-8735-06d720d6079c" containerName="init" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.724177 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a13077-0749-4821-8735-06d720d6079c" containerName="dnsmasq-dns" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.724905 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.736223 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rtkk7"] Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.745855 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db92-account-create-update-vgcjb"] Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.747437 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.756888 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.779593 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db92-account-create-update-vgcjb"] Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.852235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tjcn\" (UniqueName: \"kubernetes.io/projected/e5f4b46e-0c7d-4390-9452-d462f1569eeb-kube-api-access-4tjcn\") pod \"cinder-db-create-rtkk7\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.852281 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3f27a8-f563-4c6a-97cc-e595247ea87a-operator-scripts\") pod \"cinder-db92-account-create-update-vgcjb\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.852494 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4b46e-0c7d-4390-9452-d462f1569eeb-operator-scripts\") pod \"cinder-db-create-rtkk7\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.852964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct74v\" (UniqueName: \"kubernetes.io/projected/ff3f27a8-f563-4c6a-97cc-e595247ea87a-kube-api-access-ct74v\") pod \"cinder-db92-account-create-update-vgcjb\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.954353 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tjcn\" (UniqueName: \"kubernetes.io/projected/e5f4b46e-0c7d-4390-9452-d462f1569eeb-kube-api-access-4tjcn\") pod \"cinder-db-create-rtkk7\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.954401 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3f27a8-f563-4c6a-97cc-e595247ea87a-operator-scripts\") pod \"cinder-db92-account-create-update-vgcjb\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.954439 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4b46e-0c7d-4390-9452-d462f1569eeb-operator-scripts\") pod \"cinder-db-create-rtkk7\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.954497 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct74v\" (UniqueName: \"kubernetes.io/projected/ff3f27a8-f563-4c6a-97cc-e595247ea87a-kube-api-access-ct74v\") pod \"cinder-db92-account-create-update-vgcjb\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.955477 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3f27a8-f563-4c6a-97cc-e595247ea87a-operator-scripts\") pod \"cinder-db92-account-create-update-vgcjb\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.955488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4b46e-0c7d-4390-9452-d462f1569eeb-operator-scripts\") pod \"cinder-db-create-rtkk7\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.974278 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tjcn\" (UniqueName: \"kubernetes.io/projected/e5f4b46e-0c7d-4390-9452-d462f1569eeb-kube-api-access-4tjcn\") pod \"cinder-db-create-rtkk7\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:49 crc kubenswrapper[4758]: I1203 18:34:49.974280 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct74v\" (UniqueName: \"kubernetes.io/projected/ff3f27a8-f563-4c6a-97cc-e595247ea87a-kube-api-access-ct74v\") pod \"cinder-db92-account-create-update-vgcjb\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:50 crc kubenswrapper[4758]: I1203 18:34:50.055959 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:50 crc kubenswrapper[4758]: I1203 18:34:50.071087 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:50 crc kubenswrapper[4758]: I1203 18:34:50.587478 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rtkk7"] Dec 03 18:34:50 crc kubenswrapper[4758]: I1203 18:34:50.595443 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db92-account-create-update-vgcjb"] Dec 03 18:34:50 crc kubenswrapper[4758]: W1203 18:34:50.597069 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff3f27a8_f563_4c6a_97cc_e595247ea87a.slice/crio-4d0832db9d3586bd040df8a9dc1f20c86478bf72dcf87d2f4841feba4c642c5f WatchSource:0}: Error finding container 4d0832db9d3586bd040df8a9dc1f20c86478bf72dcf87d2f4841feba4c642c5f: Status 404 returned error can't find the container with id 4d0832db9d3586bd040df8a9dc1f20c86478bf72dcf87d2f4841feba4c642c5f Dec 03 18:34:51 crc kubenswrapper[4758]: I1203 18:34:51.556702 4758 generic.go:334] "Generic (PLEG): container finished" podID="e5f4b46e-0c7d-4390-9452-d462f1569eeb" containerID="a27b35c5d641bdee69e2177ae789fb9d457bc213b353c95777a7bdbc0b098ee1" exitCode=0 Dec 03 18:34:51 crc kubenswrapper[4758]: I1203 18:34:51.556769 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rtkk7" event={"ID":"e5f4b46e-0c7d-4390-9452-d462f1569eeb","Type":"ContainerDied","Data":"a27b35c5d641bdee69e2177ae789fb9d457bc213b353c95777a7bdbc0b098ee1"} Dec 03 18:34:51 crc kubenswrapper[4758]: I1203 18:34:51.557068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rtkk7" event={"ID":"e5f4b46e-0c7d-4390-9452-d462f1569eeb","Type":"ContainerStarted","Data":"5cfff095b664be22f5287e26309872300bb1f0de953c51627c5684cfedc00944"} Dec 03 18:34:51 crc kubenswrapper[4758]: I1203 18:34:51.560222 4758 generic.go:334] "Generic (PLEG): container finished" podID="ff3f27a8-f563-4c6a-97cc-e595247ea87a" containerID="208dbdd97f246ae0c4ea5387fda7ca671a7a1e806d73c0687a262aa6b5fbeda5" exitCode=0 Dec 03 18:34:51 crc kubenswrapper[4758]: I1203 18:34:51.560263 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db92-account-create-update-vgcjb" event={"ID":"ff3f27a8-f563-4c6a-97cc-e595247ea87a","Type":"ContainerDied","Data":"208dbdd97f246ae0c4ea5387fda7ca671a7a1e806d73c0687a262aa6b5fbeda5"} Dec 03 18:34:51 crc kubenswrapper[4758]: I1203 18:34:51.560288 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db92-account-create-update-vgcjb" event={"ID":"ff3f27a8-f563-4c6a-97cc-e595247ea87a","Type":"ContainerStarted","Data":"4d0832db9d3586bd040df8a9dc1f20c86478bf72dcf87d2f4841feba4c642c5f"} Dec 03 18:34:52 crc kubenswrapper[4758]: I1203 18:34:52.114644 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:34:52 crc kubenswrapper[4758]: E1203 18:34:52.115006 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.024996 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.032841 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.127437 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3f27a8-f563-4c6a-97cc-e595247ea87a-operator-scripts\") pod \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.127536 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4b46e-0c7d-4390-9452-d462f1569eeb-operator-scripts\") pod \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.127606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct74v\" (UniqueName: \"kubernetes.io/projected/ff3f27a8-f563-4c6a-97cc-e595247ea87a-kube-api-access-ct74v\") pod \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\" (UID: \"ff3f27a8-f563-4c6a-97cc-e595247ea87a\") " Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.127653 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tjcn\" (UniqueName: \"kubernetes.io/projected/e5f4b46e-0c7d-4390-9452-d462f1569eeb-kube-api-access-4tjcn\") pod \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\" (UID: \"e5f4b46e-0c7d-4390-9452-d462f1569eeb\") " Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.128880 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f4b46e-0c7d-4390-9452-d462f1569eeb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5f4b46e-0c7d-4390-9452-d462f1569eeb" (UID: "e5f4b46e-0c7d-4390-9452-d462f1569eeb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.129040 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3f27a8-f563-4c6a-97cc-e595247ea87a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff3f27a8-f563-4c6a-97cc-e595247ea87a" (UID: "ff3f27a8-f563-4c6a-97cc-e595247ea87a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.132924 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3f27a8-f563-4c6a-97cc-e595247ea87a-kube-api-access-ct74v" (OuterVolumeSpecName: "kube-api-access-ct74v") pod "ff3f27a8-f563-4c6a-97cc-e595247ea87a" (UID: "ff3f27a8-f563-4c6a-97cc-e595247ea87a"). InnerVolumeSpecName "kube-api-access-ct74v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.133029 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f4b46e-0c7d-4390-9452-d462f1569eeb-kube-api-access-4tjcn" (OuterVolumeSpecName: "kube-api-access-4tjcn") pod "e5f4b46e-0c7d-4390-9452-d462f1569eeb" (UID: "e5f4b46e-0c7d-4390-9452-d462f1569eeb"). InnerVolumeSpecName "kube-api-access-4tjcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.229452 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3f27a8-f563-4c6a-97cc-e595247ea87a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.229748 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4b46e-0c7d-4390-9452-d462f1569eeb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.229879 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct74v\" (UniqueName: \"kubernetes.io/projected/ff3f27a8-f563-4c6a-97cc-e595247ea87a-kube-api-access-ct74v\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.230124 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tjcn\" (UniqueName: \"kubernetes.io/projected/e5f4b46e-0c7d-4390-9452-d462f1569eeb-kube-api-access-4tjcn\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.578614 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rtkk7" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.578623 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rtkk7" event={"ID":"e5f4b46e-0c7d-4390-9452-d462f1569eeb","Type":"ContainerDied","Data":"5cfff095b664be22f5287e26309872300bb1f0de953c51627c5684cfedc00944"} Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.578708 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cfff095b664be22f5287e26309872300bb1f0de953c51627c5684cfedc00944" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.580832 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db92-account-create-update-vgcjb" event={"ID":"ff3f27a8-f563-4c6a-97cc-e595247ea87a","Type":"ContainerDied","Data":"4d0832db9d3586bd040df8a9dc1f20c86478bf72dcf87d2f4841feba4c642c5f"} Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.580876 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db92-account-create-update-vgcjb" Dec 03 18:34:53 crc kubenswrapper[4758]: I1203 18:34:53.580881 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d0832db9d3586bd040df8a9dc1f20c86478bf72dcf87d2f4841feba4c642c5f" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.972174 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xqv27"] Dec 03 18:34:54 crc kubenswrapper[4758]: E1203 18:34:54.972814 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3f27a8-f563-4c6a-97cc-e595247ea87a" containerName="mariadb-account-create-update" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.972826 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3f27a8-f563-4c6a-97cc-e595247ea87a" containerName="mariadb-account-create-update" Dec 03 18:34:54 crc kubenswrapper[4758]: E1203 18:34:54.972859 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f4b46e-0c7d-4390-9452-d462f1569eeb" containerName="mariadb-database-create" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.972865 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f4b46e-0c7d-4390-9452-d462f1569eeb" containerName="mariadb-database-create" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.973046 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3f27a8-f563-4c6a-97cc-e595247ea87a" containerName="mariadb-account-create-update" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.973061 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f4b46e-0c7d-4390-9452-d462f1569eeb" containerName="mariadb-database-create" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.973661 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.976895 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qpfd6" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.976945 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.981058 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 18:34:54 crc kubenswrapper[4758]: I1203 18:34:54.987297 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xqv27"] Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.064914 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkp5v\" (UniqueName: \"kubernetes.io/projected/b5666241-ad1d-4826-acce-2f154cbb8d6b-kube-api-access-vkp5v\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.064963 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-combined-ca-bundle\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.065004 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-scripts\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.065187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-config-data\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.065237 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-db-sync-config-data\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.065512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5666241-ad1d-4826-acce-2f154cbb8d6b-etc-machine-id\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167185 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkp5v\" (UniqueName: \"kubernetes.io/projected/b5666241-ad1d-4826-acce-2f154cbb8d6b-kube-api-access-vkp5v\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167264 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-combined-ca-bundle\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167315 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-scripts\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167441 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-config-data\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-db-sync-config-data\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167841 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5666241-ad1d-4826-acce-2f154cbb8d6b-etc-machine-id\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.167961 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5666241-ad1d-4826-acce-2f154cbb8d6b-etc-machine-id\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.172950 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-db-sync-config-data\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.173379 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-config-data\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.173507 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-combined-ca-bundle\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.180461 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-scripts\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.182782 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkp5v\" (UniqueName: \"kubernetes.io/projected/b5666241-ad1d-4826-acce-2f154cbb8d6b-kube-api-access-vkp5v\") pod \"cinder-db-sync-xqv27\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.303265 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xqv27" Dec 03 18:34:55 crc kubenswrapper[4758]: W1203 18:34:55.760129 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5666241_ad1d_4826_acce_2f154cbb8d6b.slice/crio-e7f94776f1decb366a4f23273f591852f4bcc7b628bac2815ca4e61be4fcbd43 WatchSource:0}: Error finding container e7f94776f1decb366a4f23273f591852f4bcc7b628bac2815ca4e61be4fcbd43: Status 404 returned error can't find the container with id e7f94776f1decb366a4f23273f591852f4bcc7b628bac2815ca4e61be4fcbd43 Dec 03 18:34:55 crc kubenswrapper[4758]: I1203 18:34:55.762529 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xqv27"] Dec 03 18:34:56 crc kubenswrapper[4758]: I1203 18:34:56.615601 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xqv27" event={"ID":"b5666241-ad1d-4826-acce-2f154cbb8d6b","Type":"ContainerStarted","Data":"519680d41f550cf998b5d4e5487c6d0f2d5152a7019fcb8ac684b66b1032ad80"} Dec 03 18:34:56 crc kubenswrapper[4758]: I1203 18:34:56.616149 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xqv27" event={"ID":"b5666241-ad1d-4826-acce-2f154cbb8d6b","Type":"ContainerStarted","Data":"e7f94776f1decb366a4f23273f591852f4bcc7b628bac2815ca4e61be4fcbd43"} Dec 03 18:34:56 crc kubenswrapper[4758]: I1203 18:34:56.656953 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xqv27" podStartSLOduration=2.656926247 podStartE2EDuration="2.656926247s" podCreationTimestamp="2025-12-03 18:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:34:56.6491933 +0000 UTC m=+5951.850570171" watchObservedRunningTime="2025-12-03 18:34:56.656926247 +0000 UTC m=+5951.858303108" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.226110 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6mmg7"] Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.228315 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.240344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mmg7"] Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.321380 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v7fr\" (UniqueName: \"kubernetes.io/projected/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-kube-api-access-4v7fr\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.321451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-utilities\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.321487 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-catalog-content\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.422655 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v7fr\" (UniqueName: \"kubernetes.io/projected/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-kube-api-access-4v7fr\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.422750 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-utilities\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.422785 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-catalog-content\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.423279 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-utilities\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.423330 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-catalog-content\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.440224 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v7fr\" (UniqueName: \"kubernetes.io/projected/aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24-kube-api-access-4v7fr\") pod \"redhat-operators-6mmg7\" (UID: \"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24\") " pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:58 crc kubenswrapper[4758]: I1203 18:34:58.566547 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:34:59 crc kubenswrapper[4758]: I1203 18:34:59.078144 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mmg7"] Dec 03 18:34:59 crc kubenswrapper[4758]: W1203 18:34:59.082656 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa3a0dc1_e4ad_41c2_b2c3_45c1831d4e24.slice/crio-a61a4f86464ef6d9727a486f7e6be70eecbe710ad6dcc79a4a10890e8026b287 WatchSource:0}: Error finding container a61a4f86464ef6d9727a486f7e6be70eecbe710ad6dcc79a4a10890e8026b287: Status 404 returned error can't find the container with id a61a4f86464ef6d9727a486f7e6be70eecbe710ad6dcc79a4a10890e8026b287 Dec 03 18:34:59 crc kubenswrapper[4758]: I1203 18:34:59.641265 4758 generic.go:334] "Generic (PLEG): container finished" podID="aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24" containerID="682fca86bd191003cb801ddc1ce0b0f944a2935b047ae014d29e1ebc0c335fc5" exitCode=0 Dec 03 18:34:59 crc kubenswrapper[4758]: I1203 18:34:59.641310 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmg7" event={"ID":"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24","Type":"ContainerDied","Data":"682fca86bd191003cb801ddc1ce0b0f944a2935b047ae014d29e1ebc0c335fc5"} Dec 03 18:34:59 crc kubenswrapper[4758]: I1203 18:34:59.641365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmg7" event={"ID":"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24","Type":"ContainerStarted","Data":"a61a4f86464ef6d9727a486f7e6be70eecbe710ad6dcc79a4a10890e8026b287"} Dec 03 18:35:00 crc kubenswrapper[4758]: I1203 18:35:00.651109 4758 generic.go:334] "Generic (PLEG): container finished" podID="b5666241-ad1d-4826-acce-2f154cbb8d6b" containerID="519680d41f550cf998b5d4e5487c6d0f2d5152a7019fcb8ac684b66b1032ad80" exitCode=0 Dec 03 18:35:00 crc kubenswrapper[4758]: I1203 18:35:00.651372 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xqv27" event={"ID":"b5666241-ad1d-4826-acce-2f154cbb8d6b","Type":"ContainerDied","Data":"519680d41f550cf998b5d4e5487c6d0f2d5152a7019fcb8ac684b66b1032ad80"} Dec 03 18:35:01 crc kubenswrapper[4758]: I1203 18:35:01.983762 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xqv27" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.096801 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-config-data\") pod \"b5666241-ad1d-4826-acce-2f154cbb8d6b\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.096885 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-db-sync-config-data\") pod \"b5666241-ad1d-4826-acce-2f154cbb8d6b\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.096926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-combined-ca-bundle\") pod \"b5666241-ad1d-4826-acce-2f154cbb8d6b\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.097064 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkp5v\" (UniqueName: \"kubernetes.io/projected/b5666241-ad1d-4826-acce-2f154cbb8d6b-kube-api-access-vkp5v\") pod \"b5666241-ad1d-4826-acce-2f154cbb8d6b\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.097158 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-scripts\") pod \"b5666241-ad1d-4826-acce-2f154cbb8d6b\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.097784 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5666241-ad1d-4826-acce-2f154cbb8d6b-etc-machine-id\") pod \"b5666241-ad1d-4826-acce-2f154cbb8d6b\" (UID: \"b5666241-ad1d-4826-acce-2f154cbb8d6b\") " Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.097840 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5666241-ad1d-4826-acce-2f154cbb8d6b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b5666241-ad1d-4826-acce-2f154cbb8d6b" (UID: "b5666241-ad1d-4826-acce-2f154cbb8d6b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.098178 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5666241-ad1d-4826-acce-2f154cbb8d6b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.102389 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5666241-ad1d-4826-acce-2f154cbb8d6b-kube-api-access-vkp5v" (OuterVolumeSpecName: "kube-api-access-vkp5v") pod "b5666241-ad1d-4826-acce-2f154cbb8d6b" (UID: "b5666241-ad1d-4826-acce-2f154cbb8d6b"). InnerVolumeSpecName "kube-api-access-vkp5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.102268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b5666241-ad1d-4826-acce-2f154cbb8d6b" (UID: "b5666241-ad1d-4826-acce-2f154cbb8d6b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.121286 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-scripts" (OuterVolumeSpecName: "scripts") pod "b5666241-ad1d-4826-acce-2f154cbb8d6b" (UID: "b5666241-ad1d-4826-acce-2f154cbb8d6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.123947 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5666241-ad1d-4826-acce-2f154cbb8d6b" (UID: "b5666241-ad1d-4826-acce-2f154cbb8d6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.149093 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-config-data" (OuterVolumeSpecName: "config-data") pod "b5666241-ad1d-4826-acce-2f154cbb8d6b" (UID: "b5666241-ad1d-4826-acce-2f154cbb8d6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.200153 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.200195 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.200210 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.200223 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5666241-ad1d-4826-acce-2f154cbb8d6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.200235 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkp5v\" (UniqueName: \"kubernetes.io/projected/b5666241-ad1d-4826-acce-2f154cbb8d6b-kube-api-access-vkp5v\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.670947 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xqv27" event={"ID":"b5666241-ad1d-4826-acce-2f154cbb8d6b","Type":"ContainerDied","Data":"e7f94776f1decb366a4f23273f591852f4bcc7b628bac2815ca4e61be4fcbd43"} Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.670988 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7f94776f1decb366a4f23273f591852f4bcc7b628bac2815ca4e61be4fcbd43" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.671048 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xqv27" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.974695 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7446b96cf-q5w56"] Dec 03 18:35:02 crc kubenswrapper[4758]: E1203 18:35:02.975192 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5666241-ad1d-4826-acce-2f154cbb8d6b" containerName="cinder-db-sync" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.975213 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5666241-ad1d-4826-acce-2f154cbb8d6b" containerName="cinder-db-sync" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.975401 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5666241-ad1d-4826-acce-2f154cbb8d6b" containerName="cinder-db-sync" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.977746 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:02 crc kubenswrapper[4758]: I1203 18:35:02.988758 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7446b96cf-q5w56"] Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.119586 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-dns-svc\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.119633 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlsgp\" (UniqueName: \"kubernetes.io/projected/7a67effd-e5a5-4942-ae89-c717111fe61d-kube-api-access-jlsgp\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.119699 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-config\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.119734 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-nb\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.119766 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-sb\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.165834 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.167403 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.174181 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.174472 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.174620 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qpfd6" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.176123 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.182355 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.222736 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-config\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.222814 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-nb\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.222854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-sb\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.222903 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-dns-svc\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.222933 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlsgp\" (UniqueName: \"kubernetes.io/projected/7a67effd-e5a5-4942-ae89-c717111fe61d-kube-api-access-jlsgp\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.224017 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-config\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.225073 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-nb\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.225290 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-sb\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.225585 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-dns-svc\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.251385 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlsgp\" (UniqueName: \"kubernetes.io/projected/7a67effd-e5a5-4942-ae89-c717111fe61d-kube-api-access-jlsgp\") pod \"dnsmasq-dns-7446b96cf-q5w56\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.304826 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324753 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c942e361-c212-420a-903a-e6977f8fb011-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data-custom\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324837 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ttf2\" (UniqueName: \"kubernetes.io/projected/c942e361-c212-420a-903a-e6977f8fb011-kube-api-access-7ttf2\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324871 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-scripts\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324890 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c942e361-c212-420a-903a-e6977f8fb011-logs\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324914 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.324972 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426112 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c942e361-c212-420a-903a-e6977f8fb011-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426183 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data-custom\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426208 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ttf2\" (UniqueName: \"kubernetes.io/projected/c942e361-c212-420a-903a-e6977f8fb011-kube-api-access-7ttf2\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426243 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-scripts\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426263 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c942e361-c212-420a-903a-e6977f8fb011-logs\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426290 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426353 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.426916 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c942e361-c212-420a-903a-e6977f8fb011-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.427534 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c942e361-c212-420a-903a-e6977f8fb011-logs\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.433634 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.436325 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data-custom\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.438514 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.441455 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-scripts\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.449301 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ttf2\" (UniqueName: \"kubernetes.io/projected/c942e361-c212-420a-903a-e6977f8fb011-kube-api-access-7ttf2\") pod \"cinder-api-0\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.491719 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 18:35:03 crc kubenswrapper[4758]: I1203 18:35:03.825479 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7446b96cf-q5w56"] Dec 03 18:35:03 crc kubenswrapper[4758]: W1203 18:35:03.830752 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a67effd_e5a5_4942_ae89_c717111fe61d.slice/crio-9827509f631626921a9ef3c03e6fe238e1b32e06f0d02a13403031c881f86b65 WatchSource:0}: Error finding container 9827509f631626921a9ef3c03e6fe238e1b32e06f0d02a13403031c881f86b65: Status 404 returned error can't find the container with id 9827509f631626921a9ef3c03e6fe238e1b32e06f0d02a13403031c881f86b65 Dec 03 18:35:04 crc kubenswrapper[4758]: I1203 18:35:04.060373 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:04 crc kubenswrapper[4758]: I1203 18:35:04.697261 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerID="9c2d3cf728b689802439f5a17fc863ef657bf5d9af0d84183e8641fce0557714" exitCode=0 Dec 03 18:35:04 crc kubenswrapper[4758]: I1203 18:35:04.697383 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" event={"ID":"7a67effd-e5a5-4942-ae89-c717111fe61d","Type":"ContainerDied","Data":"9c2d3cf728b689802439f5a17fc863ef657bf5d9af0d84183e8641fce0557714"} Dec 03 18:35:04 crc kubenswrapper[4758]: I1203 18:35:04.697605 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" event={"ID":"7a67effd-e5a5-4942-ae89-c717111fe61d","Type":"ContainerStarted","Data":"9827509f631626921a9ef3c03e6fe238e1b32e06f0d02a13403031c881f86b65"} Dec 03 18:35:04 crc kubenswrapper[4758]: I1203 18:35:04.700644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c942e361-c212-420a-903a-e6977f8fb011","Type":"ContainerStarted","Data":"049e0a7dde4e8f240e2fda13a73f1f2c5e7119244db1c6acc4bd0a968aef080d"} Dec 03 18:35:05 crc kubenswrapper[4758]: I1203 18:35:05.710703 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c942e361-c212-420a-903a-e6977f8fb011","Type":"ContainerStarted","Data":"ee58edb46c1472722256f05d635bc1ee927748b55b15a59dd645861e18a9ed0a"} Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.115080 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:35:06 crc kubenswrapper[4758]: E1203 18:35:06.115710 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.817051 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.817758 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="941980fe-3b3a-4183-bbbe-08a7d71563a6" containerName="nova-scheduler-scheduler" containerID="cri-o://6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3" gracePeriod=30 Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.833525 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.833928 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-log" containerID="cri-o://22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28" gracePeriod=30 Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.834467 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-api" containerID="cri-o://ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9" gracePeriod=30 Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.856967 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.857263 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerName="nova-cell0-conductor-conductor" containerID="cri-o://ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" gracePeriod=30 Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.875293 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.875538 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="f4e3e6b7-8244-48b9-9947-2547938d5206" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://cfb70d3b0fe2005455a60707770bdb91f5e52dd5a9ac0c7c57b2670fcc375fcc" gracePeriod=30 Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.897599 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.898043 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-log" containerID="cri-o://c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455" gracePeriod=30 Dec 03 18:35:06 crc kubenswrapper[4758]: I1203 18:35:06.898091 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-metadata" containerID="cri-o://37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734" gracePeriod=30 Dec 03 18:35:07 crc kubenswrapper[4758]: E1203 18:35:07.104840 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:07 crc kubenswrapper[4758]: E1203 18:35:07.106336 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:07 crc kubenswrapper[4758]: E1203 18:35:07.107528 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:07 crc kubenswrapper[4758]: E1203 18:35:07.107561 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerName="nova-cell0-conductor-conductor" Dec 03 18:35:07 crc kubenswrapper[4758]: I1203 18:35:07.750155 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4e3e6b7-8244-48b9-9947-2547938d5206" containerID="cfb70d3b0fe2005455a60707770bdb91f5e52dd5a9ac0c7c57b2670fcc375fcc" exitCode=0 Dec 03 18:35:07 crc kubenswrapper[4758]: I1203 18:35:07.750229 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f4e3e6b7-8244-48b9-9947-2547938d5206","Type":"ContainerDied","Data":"cfb70d3b0fe2005455a60707770bdb91f5e52dd5a9ac0c7c57b2670fcc375fcc"} Dec 03 18:35:07 crc kubenswrapper[4758]: I1203 18:35:07.752399 4758 generic.go:334] "Generic (PLEG): container finished" podID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerID="c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455" exitCode=143 Dec 03 18:35:07 crc kubenswrapper[4758]: I1203 18:35:07.752463 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2d83593-6b04-4aeb-95b9-a0aa0de618eb","Type":"ContainerDied","Data":"c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455"} Dec 03 18:35:07 crc kubenswrapper[4758]: I1203 18:35:07.755742 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a192022-6125-4f06-aefc-b65c68c931d0" containerID="22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28" exitCode=143 Dec 03 18:35:07 crc kubenswrapper[4758]: I1203 18:35:07.755818 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a192022-6125-4f06-aefc-b65c68c931d0","Type":"ContainerDied","Data":"22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28"} Dec 03 18:35:08 crc kubenswrapper[4758]: I1203 18:35:08.375181 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="f4e3e6b7-8244-48b9-9947-2547938d5206" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.66:6080/vnc_lite.html\": dial tcp 10.217.1.66:6080: connect: connection refused" Dec 03 18:35:09 crc kubenswrapper[4758]: I1203 18:35:09.798486 4758 generic.go:334] "Generic (PLEG): container finished" podID="941980fe-3b3a-4183-bbbe-08a7d71563a6" containerID="6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3" exitCode=0 Dec 03 18:35:09 crc kubenswrapper[4758]: I1203 18:35:09.798668 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"941980fe-3b3a-4183-bbbe-08a7d71563a6","Type":"ContainerDied","Data":"6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.017094 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.017304 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" containerName="nova-cell1-conductor-conductor" containerID="cri-o://3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84" gracePeriod=30 Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.639450 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.657871 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3 is running failed: container process not found" containerID="6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.670607 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3 is running failed: container process not found" containerID="6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.696732 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkxfn\" (UniqueName: \"kubernetes.io/projected/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-kube-api-access-vkxfn\") pod \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.696792 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-config-data\") pod \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.696855 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-logs\") pod \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.697140 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-combined-ca-bundle\") pod \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\" (UID: \"e2d83593-6b04-4aeb-95b9-a0aa0de618eb\") " Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.699024 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3 is running failed: container process not found" containerID="6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.699087 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="941980fe-3b3a-4183-bbbe-08a7d71563a6" containerName="nova-scheduler-scheduler" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.699802 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-logs" (OuterVolumeSpecName: "logs") pod "e2d83593-6b04-4aeb-95b9-a0aa0de618eb" (UID: "e2d83593-6b04-4aeb-95b9-a0aa0de618eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.720431 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-kube-api-access-vkxfn" (OuterVolumeSpecName: "kube-api-access-vkxfn") pod "e2d83593-6b04-4aeb-95b9-a0aa0de618eb" (UID: "e2d83593-6b04-4aeb-95b9-a0aa0de618eb"). InnerVolumeSpecName "kube-api-access-vkxfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.763865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d83593-6b04-4aeb-95b9-a0aa0de618eb" (UID: "e2d83593-6b04-4aeb-95b9-a0aa0de618eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.779286 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-config-data" (OuterVolumeSpecName: "config-data") pod "e2d83593-6b04-4aeb-95b9-a0aa0de618eb" (UID: "e2d83593-6b04-4aeb-95b9-a0aa0de618eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.781075 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.787736 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.807762 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-combined-ca-bundle\") pod \"941980fe-3b3a-4183-bbbe-08a7d71563a6\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808004 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxdj2\" (UniqueName: \"kubernetes.io/projected/f4e3e6b7-8244-48b9-9947-2547938d5206-kube-api-access-kxdj2\") pod \"f4e3e6b7-8244-48b9-9947-2547938d5206\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808087 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htl54\" (UniqueName: \"kubernetes.io/projected/941980fe-3b3a-4183-bbbe-08a7d71563a6-kube-api-access-htl54\") pod \"941980fe-3b3a-4183-bbbe-08a7d71563a6\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808228 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-combined-ca-bundle\") pod \"f4e3e6b7-8244-48b9-9947-2547938d5206\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808264 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data\") pod \"f4e3e6b7-8244-48b9-9947-2547938d5206\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808338 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-config-data\") pod \"941980fe-3b3a-4183-bbbe-08a7d71563a6\" (UID: \"941980fe-3b3a-4183-bbbe-08a7d71563a6\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808786 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808807 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkxfn\" (UniqueName: \"kubernetes.io/projected/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-kube-api-access-vkxfn\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808820 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.808830 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d83593-6b04-4aeb-95b9-a0aa0de618eb-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.816838 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e3e6b7-8244-48b9-9947-2547938d5206-kube-api-access-kxdj2" (OuterVolumeSpecName: "kube-api-access-kxdj2") pod "f4e3e6b7-8244-48b9-9947-2547938d5206" (UID: "f4e3e6b7-8244-48b9-9947-2547938d5206"). InnerVolumeSpecName "kube-api-access-kxdj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.817005 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/941980fe-3b3a-4183-bbbe-08a7d71563a6-kube-api-access-htl54" (OuterVolumeSpecName: "kube-api-access-htl54") pod "941980fe-3b3a-4183-bbbe-08a7d71563a6" (UID: "941980fe-3b3a-4183-bbbe-08a7d71563a6"). InnerVolumeSpecName "kube-api-access-htl54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.845391 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "941980fe-3b3a-4183-bbbe-08a7d71563a6" (UID: "941980fe-3b3a-4183-bbbe-08a7d71563a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.846268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4e3e6b7-8244-48b9-9947-2547938d5206" (UID: "f4e3e6b7-8244-48b9-9947-2547938d5206"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.851429 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.851573 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.851654 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f4e3e6b7-8244-48b9-9947-2547938d5206","Type":"ContainerDied","Data":"476de057ee896558bbcfe44097869e9c14f45cdf71895de2b7e75d812d559a64"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.851728 4758 scope.go:117] "RemoveContainer" containerID="cfb70d3b0fe2005455a60707770bdb91f5e52dd5a9ac0c7c57b2670fcc375fcc" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.876851 4758 generic.go:334] "Generic (PLEG): container finished" podID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerID="37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734" exitCode=0 Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.876958 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2d83593-6b04-4aeb-95b9-a0aa0de618eb","Type":"ContainerDied","Data":"37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.876991 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2d83593-6b04-4aeb-95b9-a0aa0de618eb","Type":"ContainerDied","Data":"1cabb2cb6812518bd8419976cafa5ff5e3bbd9b68b0b9a7dc74a83bb1df22139"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.877114 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.882838 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-config-data" (OuterVolumeSpecName: "config-data") pod "941980fe-3b3a-4183-bbbe-08a7d71563a6" (UID: "941980fe-3b3a-4183-bbbe-08a7d71563a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.910550 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a192022-6125-4f06-aefc-b65c68c931d0" containerID="ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9" exitCode=0 Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.910795 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a192022-6125-4f06-aefc-b65c68c931d0","Type":"ContainerDied","Data":"ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.910830 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a192022-6125-4f06-aefc-b65c68c931d0","Type":"ContainerDied","Data":"e67cd50eb922d0c70a442eaeabcb4d760d9283317975d19af60c65a6e52991ea"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.910836 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.911588 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data" (OuterVolumeSpecName: "config-data") pod "f4e3e6b7-8244-48b9-9947-2547938d5206" (UID: "f4e3e6b7-8244-48b9-9947-2547938d5206"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.912046 4758 scope.go:117] "RemoveContainer" containerID="37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.912933 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data\") pod \"f4e3e6b7-8244-48b9-9947-2547938d5206\" (UID: \"f4e3e6b7-8244-48b9-9947-2547938d5206\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.913164 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2r2b\" (UniqueName: \"kubernetes.io/projected/3a192022-6125-4f06-aefc-b65c68c931d0-kube-api-access-z2r2b\") pod \"3a192022-6125-4f06-aefc-b65c68c931d0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " Dec 03 18:35:10 crc kubenswrapper[4758]: W1203 18:35:10.913218 4758 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f4e3e6b7-8244-48b9-9947-2547938d5206/volumes/kubernetes.io~secret/config-data Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.913248 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data" (OuterVolumeSpecName: "config-data") pod "f4e3e6b7-8244-48b9-9947-2547938d5206" (UID: "f4e3e6b7-8244-48b9-9947-2547938d5206"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.913270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-config-data\") pod \"3a192022-6125-4f06-aefc-b65c68c931d0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.913363 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a192022-6125-4f06-aefc-b65c68c931d0-logs\") pod \"3a192022-6125-4f06-aefc-b65c68c931d0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914138 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a192022-6125-4f06-aefc-b65c68c931d0-logs" (OuterVolumeSpecName: "logs") pod "3a192022-6125-4f06-aefc-b65c68c931d0" (UID: "3a192022-6125-4f06-aefc-b65c68c931d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914292 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914322 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxdj2\" (UniqueName: \"kubernetes.io/projected/f4e3e6b7-8244-48b9-9947-2547938d5206-kube-api-access-kxdj2\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914336 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a192022-6125-4f06-aefc-b65c68c931d0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914351 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htl54\" (UniqueName: \"kubernetes.io/projected/941980fe-3b3a-4183-bbbe-08a7d71563a6-kube-api-access-htl54\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914361 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914370 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e3e6b7-8244-48b9-9947-2547938d5206-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.914378 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/941980fe-3b3a-4183-bbbe-08a7d71563a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.925090 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"941980fe-3b3a-4183-bbbe-08a7d71563a6","Type":"ContainerDied","Data":"dd4fe19eb80298348193fc2d12e4ea2a172cf315ffda6a1778f5699b409db7a5"} Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.925203 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.925492 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a192022-6125-4f06-aefc-b65c68c931d0-kube-api-access-z2r2b" (OuterVolumeSpecName: "kube-api-access-z2r2b") pod "3a192022-6125-4f06-aefc-b65c68c931d0" (UID: "3a192022-6125-4f06-aefc-b65c68c931d0"). InnerVolumeSpecName "kube-api-access-z2r2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.947153 4758 scope.go:117] "RemoveContainer" containerID="c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.964794 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.975652 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.983640 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.984104 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e3e6b7-8244-48b9-9947-2547938d5206" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984116 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e3e6b7-8244-48b9-9947-2547938d5206" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.984132 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941980fe-3b3a-4183-bbbe-08a7d71563a6" containerName="nova-scheduler-scheduler" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984140 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="941980fe-3b3a-4183-bbbe-08a7d71563a6" containerName="nova-scheduler-scheduler" Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.984184 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-api" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984192 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-api" Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.984206 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-metadata" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984212 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-metadata" Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.984231 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-log" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984236 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-log" Dec 03 18:35:10 crc kubenswrapper[4758]: E1203 18:35:10.984249 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-log" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984255 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-log" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984442 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-metadata" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984458 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="941980fe-3b3a-4183-bbbe-08a7d71563a6" containerName="nova-scheduler-scheduler" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984467 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e3e6b7-8244-48b9-9947-2547938d5206" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984476 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-log" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984489 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" containerName="nova-metadata-log" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.984503 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" containerName="nova-api-api" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.985503 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.991337 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 18:35:10 crc kubenswrapper[4758]: I1203 18:35:10.994107 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.004212 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.013242 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.015286 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-combined-ca-bundle\") pod \"3a192022-6125-4f06-aefc-b65c68c931d0\" (UID: \"3a192022-6125-4f06-aefc-b65c68c931d0\") " Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.015561 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb17f73-7b5f-4225-829b-53b13343ea77-logs\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.015604 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.015669 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjwb4\" (UniqueName: \"kubernetes.io/projected/ffb17f73-7b5f-4225-829b-53b13343ea77-kube-api-access-kjwb4\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.015726 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-config-data\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.015837 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2r2b\" (UniqueName: \"kubernetes.io/projected/3a192022-6125-4f06-aefc-b65c68c931d0-kube-api-access-z2r2b\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.016865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-config-data" (OuterVolumeSpecName: "config-data") pod "3a192022-6125-4f06-aefc-b65c68c931d0" (UID: "3a192022-6125-4f06-aefc-b65c68c931d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.020828 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.023337 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.026375 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.037791 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.041971 4758 scope.go:117] "RemoveContainer" containerID="37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734" Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.069063 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734\": container with ID starting with 37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734 not found: ID does not exist" containerID="37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.069109 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734"} err="failed to get container status \"37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734\": rpc error: code = NotFound desc = could not find container \"37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734\": container with ID starting with 37ef1ef7b698225e9e79a5ee41977b5a914d2bddf5e083e0e355fb9eff905734 not found: ID does not exist" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.069135 4758 scope.go:117] "RemoveContainer" containerID="c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.072170 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a192022-6125-4f06-aefc-b65c68c931d0" (UID: "3a192022-6125-4f06-aefc-b65c68c931d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.074013 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455\": container with ID starting with c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455 not found: ID does not exist" containerID="c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.074076 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455"} err="failed to get container status \"c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455\": rpc error: code = NotFound desc = could not find container \"c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455\": container with ID starting with c0c020b215b2a48040f0d0c1f82788b18b2f130be6ae413a29776941e8798455 not found: ID does not exist" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.074109 4758 scope.go:117] "RemoveContainer" containerID="ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.117914 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb17f73-7b5f-4225-829b-53b13343ea77-logs\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.117971 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.118826 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb17f73-7b5f-4225-829b-53b13343ea77-logs\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.119517 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjwb4\" (UniqueName: \"kubernetes.io/projected/ffb17f73-7b5f-4225-829b-53b13343ea77-kube-api-access-kjwb4\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.119601 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-config-data\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.119726 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.119750 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a192022-6125-4f06-aefc-b65c68c931d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.123971 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.127304 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-config-data\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.149708 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="941980fe-3b3a-4183-bbbe-08a7d71563a6" path="/var/lib/kubelet/pods/941980fe-3b3a-4183-bbbe-08a7d71563a6/volumes" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.150349 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d83593-6b04-4aeb-95b9-a0aa0de618eb" path="/var/lib/kubelet/pods/e2d83593-6b04-4aeb-95b9-a0aa0de618eb/volumes" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.154395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjwb4\" (UniqueName: \"kubernetes.io/projected/ffb17f73-7b5f-4225-829b-53b13343ea77-kube-api-access-kjwb4\") pod \"nova-metadata-0\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.221503 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4zkq\" (UniqueName: \"kubernetes.io/projected/58aed963-c66b-445d-8c0f-28dc1c0694f8-kube-api-access-h4zkq\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.221960 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.222050 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-config-data\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.294906 4758 scope.go:117] "RemoveContainer" containerID="22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.327209 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.327411 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-config-data\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.328479 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4zkq\" (UniqueName: \"kubernetes.io/projected/58aed963-c66b-445d-8c0f-28dc1c0694f8-kube-api-access-h4zkq\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.332052 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.334844 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.336297 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84 is running failed: container process not found" containerID="3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.336672 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84 is running failed: container process not found" containerID="3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.338345 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-config-data\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.338560 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84 is running failed: container process not found" containerID="3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.338606 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" containerName="nova-cell1-conductor-conductor" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.340521 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.357374 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4zkq\" (UniqueName: \"kubernetes.io/projected/58aed963-c66b-445d-8c0f-28dc1c0694f8-kube-api-access-h4zkq\") pod \"nova-scheduler-0\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.364383 4758 scope.go:117] "RemoveContainer" containerID="ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9" Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.365248 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9\": container with ID starting with ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9 not found: ID does not exist" containerID="ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.365293 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9"} err="failed to get container status \"ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9\": rpc error: code = NotFound desc = could not find container \"ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9\": container with ID starting with ef2ede923d57b7a568a402fa6556c8239f6bd496830df12487be3d410eef8fc9 not found: ID does not exist" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.365319 4758 scope.go:117] "RemoveContainer" containerID="22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28" Dec 03 18:35:11 crc kubenswrapper[4758]: E1203 18:35:11.372712 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28\": container with ID starting with 22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28 not found: ID does not exist" containerID="22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.372754 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28"} err="failed to get container status \"22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28\": rpc error: code = NotFound desc = could not find container \"22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28\": container with ID starting with 22580c9dae79a7e3a8082aea1cffbe01d1c458d5ead48b2c7d2de241a0b91d28 not found: ID does not exist" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.372780 4758 scope.go:117] "RemoveContainer" containerID="6e00be8fa77063efd5e8a706f7db1008fb9954bec59adb939a3e20a24c94b1b3" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.380064 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.392135 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.403755 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.410916 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.412496 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.417448 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.425503 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.427650 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.429366 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430074 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl6vr\" (UniqueName: \"kubernetes.io/projected/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-kube-api-access-bl6vr\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3714e-e415-4d7a-8710-9050fdae8e0b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430198 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc3714e-e415-4d7a-8710-9050fdae8e0b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430220 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-logs\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430249 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-config-data\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.430267 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczw6\" (UniqueName: \"kubernetes.io/projected/dfc3714e-e415-4d7a-8710-9050fdae8e0b-kube-api-access-pczw6\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.438371 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.449653 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533000 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc3714e-e415-4d7a-8710-9050fdae8e0b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533044 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-logs\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533077 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-config-data\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533097 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczw6\" (UniqueName: \"kubernetes.io/projected/dfc3714e-e415-4d7a-8710-9050fdae8e0b-kube-api-access-pczw6\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533165 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl6vr\" (UniqueName: \"kubernetes.io/projected/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-kube-api-access-bl6vr\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3714e-e415-4d7a-8710-9050fdae8e0b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.533232 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.534272 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-logs\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.539342 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.539465 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc3714e-e415-4d7a-8710-9050fdae8e0b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.540615 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-config-data\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.541136 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3714e-e415-4d7a-8710-9050fdae8e0b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.554221 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczw6\" (UniqueName: \"kubernetes.io/projected/dfc3714e-e415-4d7a-8710-9050fdae8e0b-kube-api-access-pczw6\") pod \"nova-cell1-novncproxy-0\" (UID: \"dfc3714e-e415-4d7a-8710-9050fdae8e0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.555923 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl6vr\" (UniqueName: \"kubernetes.io/projected/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-kube-api-access-bl6vr\") pod \"nova-api-0\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.593370 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.814066 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.832964 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.929077 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.960721 4758 generic.go:334] "Generic (PLEG): container finished" podID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" containerID="3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84" exitCode=0 Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.960801 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4e03e4bb-9e94-4ff1-bc70-e74cfc176094","Type":"ContainerDied","Data":"3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84"} Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.965127 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" event={"ID":"7a67effd-e5a5-4942-ae89-c717111fe61d","Type":"ContainerStarted","Data":"e61e072ca376bf4682ba7840dcbb62a3e7f56a0947d00c64d0333e16be43e17b"} Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.965320 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.968413 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c942e361-c212-420a-903a-e6977f8fb011","Type":"ContainerStarted","Data":"a9e5ebad6b7d6ba9e1dc792a00debe6467eedb6089ee5c19f2a948d6613d6e90"} Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.968558 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.971456 4758 generic.go:334] "Generic (PLEG): container finished" podID="aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24" containerID="0049d49e55b605c3b0908f8d6c213dc275c30f71b896a23e130a6409c8e4bf3f" exitCode=0 Dec 03 18:35:11 crc kubenswrapper[4758]: I1203 18:35:11.971550 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmg7" event={"ID":"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24","Type":"ContainerDied","Data":"0049d49e55b605c3b0908f8d6c213dc275c30f71b896a23e130a6409c8e4bf3f"} Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.002132 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" podStartSLOduration=10.002107494 podStartE2EDuration="10.002107494s" podCreationTimestamp="2025-12-03 18:35:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:11.985810042 +0000 UTC m=+5967.187186913" watchObservedRunningTime="2025-12-03 18:35:12.002107494 +0000 UTC m=+5967.203484355" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.084073 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=9.084049533 podStartE2EDuration="9.084049533s" podCreationTimestamp="2025-12-03 18:35:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:12.019592997 +0000 UTC m=+5967.220969848" watchObservedRunningTime="2025-12-03 18:35:12.084049533 +0000 UTC m=+5967.285426394" Dec 03 18:35:12 crc kubenswrapper[4758]: E1203 18:35:12.132319 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:12 crc kubenswrapper[4758]: E1203 18:35:12.134363 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:12 crc kubenswrapper[4758]: E1203 18:35:12.140326 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa3a0dc1_e4ad_41c2_b2c3_45c1831d4e24.slice/crio-0049d49e55b605c3b0908f8d6c213dc275c30f71b896a23e130a6409c8e4bf3f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa3a0dc1_e4ad_41c2_b2c3_45c1831d4e24.slice/crio-conmon-0049d49e55b605c3b0908f8d6c213dc275c30f71b896a23e130a6409c8e4bf3f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 18:35:12 crc kubenswrapper[4758]: E1203 18:35:12.149579 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 18:35:12 crc kubenswrapper[4758]: E1203 18:35:12.149960 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerName="nova-cell0-conductor-conductor" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.156884 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.260180 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.357577 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnhgf\" (UniqueName: \"kubernetes.io/projected/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-kube-api-access-qnhgf\") pod \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.357728 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-config-data\") pod \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.357778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-combined-ca-bundle\") pod \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\" (UID: \"4e03e4bb-9e94-4ff1-bc70-e74cfc176094\") " Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.430867 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-kube-api-access-qnhgf" (OuterVolumeSpecName: "kube-api-access-qnhgf") pod "4e03e4bb-9e94-4ff1-bc70-e74cfc176094" (UID: "4e03e4bb-9e94-4ff1-bc70-e74cfc176094"). InnerVolumeSpecName "kube-api-access-qnhgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.438840 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e03e4bb-9e94-4ff1-bc70-e74cfc176094" (UID: "4e03e4bb-9e94-4ff1-bc70-e74cfc176094"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.447064 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-config-data" (OuterVolumeSpecName: "config-data") pod "4e03e4bb-9e94-4ff1-bc70-e74cfc176094" (UID: "4e03e4bb-9e94-4ff1-bc70-e74cfc176094"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.465867 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnhgf\" (UniqueName: \"kubernetes.io/projected/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-kube-api-access-qnhgf\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.465904 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.465919 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e03e4bb-9e94-4ff1-bc70-e74cfc176094-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.473091 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.536622 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 18:35:12 crc kubenswrapper[4758]: W1203 18:35:12.548943 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfc3714e_e415_4d7a_8710_9050fdae8e0b.slice/crio-7100c8dd00454c39ae18e751bb7b1e8e34d8cc138852cb3bd2aca9ca863d878c WatchSource:0}: Error finding container 7100c8dd00454c39ae18e751bb7b1e8e34d8cc138852cb3bd2aca9ca863d878c: Status 404 returned error can't find the container with id 7100c8dd00454c39ae18e751bb7b1e8e34d8cc138852cb3bd2aca9ca863d878c Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.986061 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ffb17f73-7b5f-4225-829b-53b13343ea77","Type":"ContainerStarted","Data":"c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4"} Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.986100 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ffb17f73-7b5f-4225-829b-53b13343ea77","Type":"ContainerStarted","Data":"7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741"} Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.986111 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ffb17f73-7b5f-4225-829b-53b13343ea77","Type":"ContainerStarted","Data":"4d38561873a64e6dc5b78e857a1bfbbc9f7a1e1a320d6cad644d5e5a597acf5d"} Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.987938 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99ebca78-76cf-4cf6-817b-b9a421d8cfd3","Type":"ContainerStarted","Data":"004d80a978ced39f198ca85260643f8bbfa0e18efa48068f072a2d9efe095e6f"} Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.987967 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99ebca78-76cf-4cf6-817b-b9a421d8cfd3","Type":"ContainerStarted","Data":"1a5432a170aa183623673c3fd33edd1efb2c079a36a1763d035a1fdeafae0bc5"} Dec 03 18:35:12 crc kubenswrapper[4758]: I1203 18:35:12.991935 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmg7" event={"ID":"aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24","Type":"ContainerStarted","Data":"94be4f1abb720b607880c9774c8efd99c72adf43756cbaa12af5fa0e47378b74"} Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.007127 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4e03e4bb-9e94-4ff1-bc70-e74cfc176094","Type":"ContainerDied","Data":"3141891fd34bb66c7ae90cffbcb9df9c50c7807263610a09a0929e119674a427"} Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.007183 4758 scope.go:117] "RemoveContainer" containerID="3fa8bd41b188c263b36b67b9ab9373b28444c440632a0ddd7533ed6f85248e84" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.007399 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.010213 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.010177945 podStartE2EDuration="3.010177945s" podCreationTimestamp="2025-12-03 18:35:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:13.003860164 +0000 UTC m=+5968.205237025" watchObservedRunningTime="2025-12-03 18:35:13.010177945 +0000 UTC m=+5968.211554806" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.016212 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58aed963-c66b-445d-8c0f-28dc1c0694f8","Type":"ContainerStarted","Data":"4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb"} Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.016264 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58aed963-c66b-445d-8c0f-28dc1c0694f8","Type":"ContainerStarted","Data":"713d10a5476970a53ad24afd1d0a1b8dbbe2d8f8d37511c8a709dc7cab34f816"} Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.022929 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dfc3714e-e415-4d7a-8710-9050fdae8e0b","Type":"ContainerStarted","Data":"1ba3360b615a80e4dc22a760de53a2a45c28baedd87a90db91ebbb61a981ea07"} Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.023253 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dfc3714e-e415-4d7a-8710-9050fdae8e0b","Type":"ContainerStarted","Data":"7100c8dd00454c39ae18e751bb7b1e8e34d8cc138852cb3bd2aca9ca863d878c"} Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.036150 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6mmg7" podStartSLOduration=2.075764834 podStartE2EDuration="15.036124078s" podCreationTimestamp="2025-12-03 18:34:58 +0000 UTC" firstStartedPulling="2025-12-03 18:34:59.642728816 +0000 UTC m=+5954.844105687" lastFinishedPulling="2025-12-03 18:35:12.60308807 +0000 UTC m=+5967.804464931" observedRunningTime="2025-12-03 18:35:13.027417312 +0000 UTC m=+5968.228794183" watchObservedRunningTime="2025-12-03 18:35:13.036124078 +0000 UTC m=+5968.237500939" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.073214 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.073192622 podStartE2EDuration="3.073192622s" podCreationTimestamp="2025-12-03 18:35:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:13.052175583 +0000 UTC m=+5968.253552454" watchObservedRunningTime="2025-12-03 18:35:13.073192622 +0000 UTC m=+5968.274569483" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.111403 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.111377366 podStartE2EDuration="2.111377366s" podCreationTimestamp="2025-12-03 18:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:13.074321973 +0000 UTC m=+5968.275698834" watchObservedRunningTime="2025-12-03 18:35:13.111377366 +0000 UTC m=+5968.312754237" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.133388 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a192022-6125-4f06-aefc-b65c68c931d0" path="/var/lib/kubelet/pods/3a192022-6125-4f06-aefc-b65c68c931d0/volumes" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.134273 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e3e6b7-8244-48b9-9947-2547938d5206" path="/var/lib/kubelet/pods/f4e3e6b7-8244-48b9-9947-2547938d5206/volumes" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.134917 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.139507 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.150736 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:35:13 crc kubenswrapper[4758]: E1203 18:35:13.151256 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" containerName="nova-cell1-conductor-conductor" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.151291 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" containerName="nova-cell1-conductor-conductor" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.151509 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" containerName="nova-cell1-conductor-conductor" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.152281 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.155387 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.161824 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.280813 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.280867 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.280913 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rbrv\" (UniqueName: \"kubernetes.io/projected/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-kube-api-access-2rbrv\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.382331 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.382368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.382405 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rbrv\" (UniqueName: \"kubernetes.io/projected/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-kube-api-access-2rbrv\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.388252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.388382 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.401226 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rbrv\" (UniqueName: \"kubernetes.io/projected/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-kube-api-access-2rbrv\") pod \"nova-cell1-conductor-0\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:13 crc kubenswrapper[4758]: I1203 18:35:13.496224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:14 crc kubenswrapper[4758]: I1203 18:35:14.048149 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 18:35:14 crc kubenswrapper[4758]: I1203 18:35:14.053195 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99ebca78-76cf-4cf6-817b-b9a421d8cfd3","Type":"ContainerStarted","Data":"707b6849d2d683a60141de7af8106e6805d937ae961eac28c6acd1631443d051"} Dec 03 18:35:14 crc kubenswrapper[4758]: I1203 18:35:14.102731 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.102707834 podStartE2EDuration="3.102707834s" podCreationTimestamp="2025-12-03 18:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:14.081634143 +0000 UTC m=+5969.283011024" watchObservedRunningTime="2025-12-03 18:35:14.102707834 +0000 UTC m=+5969.304084695" Dec 03 18:35:15 crc kubenswrapper[4758]: I1203 18:35:15.071262 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"44772cdb-c8fc-49a8-a2b7-56a81cec54ce","Type":"ContainerStarted","Data":"9eeed526487510d141effa3293138fe5c24f0a9d0b697fefc042da231282688a"} Dec 03 18:35:15 crc kubenswrapper[4758]: I1203 18:35:15.072838 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:15 crc kubenswrapper[4758]: I1203 18:35:15.072851 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"44772cdb-c8fc-49a8-a2b7-56a81cec54ce","Type":"ContainerStarted","Data":"4b09d17d5d3f96dbf428f076c5ea56a2c6cef5f6c641a87f441c70026658bf55"} Dec 03 18:35:15 crc kubenswrapper[4758]: I1203 18:35:15.096766 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.096748235 podStartE2EDuration="2.096748235s" podCreationTimestamp="2025-12-03 18:35:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:15.089599302 +0000 UTC m=+5970.290976163" watchObservedRunningTime="2025-12-03 18:35:15.096748235 +0000 UTC m=+5970.298125096" Dec 03 18:35:15 crc kubenswrapper[4758]: I1203 18:35:15.135021 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e03e4bb-9e94-4ff1-bc70-e74cfc176094" path="/var/lib/kubelet/pods/4e03e4bb-9e94-4ff1-bc70-e74cfc176094/volumes" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.081453 4758 generic.go:334] "Generic (PLEG): container finished" podID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" exitCode=0 Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.081532 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"39670374-d4c0-40d2-b1c2-d838b472fdc1","Type":"ContainerDied","Data":"ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf"} Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.082596 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"39670374-d4c0-40d2-b1c2-d838b472fdc1","Type":"ContainerDied","Data":"024ca130aaca69ce7f856fc67c3ed4b89f85b181529be99e2c8b734d695f0450"} Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.082614 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="024ca130aaca69ce7f856fc67c3ed4b89f85b181529be99e2c8b734d695f0450" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.102556 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.147495 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wznb5\" (UniqueName: \"kubernetes.io/projected/39670374-d4c0-40d2-b1c2-d838b472fdc1-kube-api-access-wznb5\") pod \"39670374-d4c0-40d2-b1c2-d838b472fdc1\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.147551 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-combined-ca-bundle\") pod \"39670374-d4c0-40d2-b1c2-d838b472fdc1\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.147728 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-config-data\") pod \"39670374-d4c0-40d2-b1c2-d838b472fdc1\" (UID: \"39670374-d4c0-40d2-b1c2-d838b472fdc1\") " Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.167023 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39670374-d4c0-40d2-b1c2-d838b472fdc1-kube-api-access-wznb5" (OuterVolumeSpecName: "kube-api-access-wznb5") pod "39670374-d4c0-40d2-b1c2-d838b472fdc1" (UID: "39670374-d4c0-40d2-b1c2-d838b472fdc1"). InnerVolumeSpecName "kube-api-access-wznb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.183961 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39670374-d4c0-40d2-b1c2-d838b472fdc1" (UID: "39670374-d4c0-40d2-b1c2-d838b472fdc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.188094 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-config-data" (OuterVolumeSpecName: "config-data") pod "39670374-d4c0-40d2-b1c2-d838b472fdc1" (UID: "39670374-d4c0-40d2-b1c2-d838b472fdc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.249652 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.249710 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wznb5\" (UniqueName: \"kubernetes.io/projected/39670374-d4c0-40d2-b1c2-d838b472fdc1-kube-api-access-wznb5\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.249722 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39670374-d4c0-40d2-b1c2-d838b472fdc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.341070 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.341179 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.594053 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 18:35:16 crc kubenswrapper[4758]: I1203 18:35:16.814610 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.091721 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.126104 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.135307 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.153527 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:35:17 crc kubenswrapper[4758]: E1203 18:35:17.154039 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerName="nova-cell0-conductor-conductor" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.154057 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerName="nova-cell0-conductor-conductor" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.154223 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" containerName="nova-cell0-conductor-conductor" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.154882 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.157390 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.183576 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.264898 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.264964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxgth\" (UniqueName: \"kubernetes.io/projected/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-kube-api-access-wxgth\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.265707 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.367192 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.367290 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxgth\" (UniqueName: \"kubernetes.io/projected/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-kube-api-access-wxgth\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.367348 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.371727 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.373592 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.387984 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxgth\" (UniqueName: \"kubernetes.io/projected/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-kube-api-access-wxgth\") pod \"nova-cell0-conductor-0\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.475974 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:17 crc kubenswrapper[4758]: W1203 18:35:17.914843 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1f5f4e9_0563_4d1a_b686_51d7657dfc76.slice/crio-29f981a03cac529ac6ea332443ce93f97696a5bc939c5b454718edeb961436e2 WatchSource:0}: Error finding container 29f981a03cac529ac6ea332443ce93f97696a5bc939c5b454718edeb961436e2: Status 404 returned error can't find the container with id 29f981a03cac529ac6ea332443ce93f97696a5bc939c5b454718edeb961436e2 Dec 03 18:35:17 crc kubenswrapper[4758]: I1203 18:35:17.919183 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.108551 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d1f5f4e9-0563-4d1a-b686-51d7657dfc76","Type":"ContainerStarted","Data":"29f981a03cac529ac6ea332443ce93f97696a5bc939c5b454718edeb961436e2"} Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.306877 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.371884 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65fb5d464f-w9xqz"] Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.372302 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerName="dnsmasq-dns" containerID="cri-o://61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266" gracePeriod=10 Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.567972 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.572849 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.641901 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:35:18 crc kubenswrapper[4758]: I1203 18:35:18.932545 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.101559 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ht2j\" (UniqueName: \"kubernetes.io/projected/21bc458f-4b56-4ed0-b3e4-fbad588d8669-kube-api-access-7ht2j\") pod \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.101662 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-nb\") pod \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.101714 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-config\") pod \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.101750 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-sb\") pod \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.101772 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-dns-svc\") pod \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\" (UID: \"21bc458f-4b56-4ed0-b3e4-fbad588d8669\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.109883 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21bc458f-4b56-4ed0-b3e4-fbad588d8669-kube-api-access-7ht2j" (OuterVolumeSpecName: "kube-api-access-7ht2j") pod "21bc458f-4b56-4ed0-b3e4-fbad588d8669" (UID: "21bc458f-4b56-4ed0-b3e4-fbad588d8669"). InnerVolumeSpecName "kube-api-access-7ht2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.127450 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39670374-d4c0-40d2-b1c2-d838b472fdc1" path="/var/lib/kubelet/pods/39670374-d4c0-40d2-b1c2-d838b472fdc1/volumes" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.139931 4758 generic.go:334] "Generic (PLEG): container finished" podID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerID="61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266" exitCode=0 Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.140586 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.140571133 podStartE2EDuration="2.140571133s" podCreationTimestamp="2025-12-03 18:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:19.137834479 +0000 UTC m=+5974.339211350" watchObservedRunningTime="2025-12-03 18:35:19.140571133 +0000 UTC m=+5974.341947994" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.141278 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.153092 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d1f5f4e9-0563-4d1a-b686-51d7657dfc76","Type":"ContainerStarted","Data":"a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d"} Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.153136 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.153149 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" event={"ID":"21bc458f-4b56-4ed0-b3e4-fbad588d8669","Type":"ContainerDied","Data":"61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266"} Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.153163 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65fb5d464f-w9xqz" event={"ID":"21bc458f-4b56-4ed0-b3e4-fbad588d8669","Type":"ContainerDied","Data":"e6e8c03b04f027e0f4a77ad77f6e7814e915d116c7fc5a8271f6c1ee4a4ae071"} Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.153181 4758 scope.go:117] "RemoveContainer" containerID="61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.154581 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-config" (OuterVolumeSpecName: "config") pod "21bc458f-4b56-4ed0-b3e4-fbad588d8669" (UID: "21bc458f-4b56-4ed0-b3e4-fbad588d8669"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.165251 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21bc458f-4b56-4ed0-b3e4-fbad588d8669" (UID: "21bc458f-4b56-4ed0-b3e4-fbad588d8669"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.171201 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21bc458f-4b56-4ed0-b3e4-fbad588d8669" (UID: "21bc458f-4b56-4ed0-b3e4-fbad588d8669"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.172140 4758 scope.go:117] "RemoveContainer" containerID="bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.187890 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21bc458f-4b56-4ed0-b3e4-fbad588d8669" (UID: "21bc458f-4b56-4ed0-b3e4-fbad588d8669"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.201741 4758 scope.go:117] "RemoveContainer" containerID="61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266" Dec 03 18:35:19 crc kubenswrapper[4758]: E1203 18:35:19.202264 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266\": container with ID starting with 61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266 not found: ID does not exist" containerID="61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.202324 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266"} err="failed to get container status \"61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266\": rpc error: code = NotFound desc = could not find container \"61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266\": container with ID starting with 61e2676dbed6288a36f3e5d128e929c521bdcdba9e81fef1b960f76159627266 not found: ID does not exist" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.202358 4758 scope.go:117] "RemoveContainer" containerID="bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654" Dec 03 18:35:19 crc kubenswrapper[4758]: E1203 18:35:19.202776 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654\": container with ID starting with bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654 not found: ID does not exist" containerID="bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.202826 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654"} err="failed to get container status \"bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654\": rpc error: code = NotFound desc = could not find container \"bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654\": container with ID starting with bafe43fea45d8ec258be71f06dbe17e113b5f35db92489927c389f25813f9654 not found: ID does not exist" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.204347 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.204366 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.204376 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.204387 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bc458f-4b56-4ed0-b3e4-fbad588d8669-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.204495 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ht2j\" (UniqueName: \"kubernetes.io/projected/21bc458f-4b56-4ed0-b3e4-fbad588d8669-kube-api-access-7ht2j\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.209511 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6mmg7" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.284874 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mmg7"] Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.339281 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m8ltr"] Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.339512 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m8ltr" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="registry-server" containerID="cri-o://7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb" gracePeriod=2 Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.590294 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65fb5d464f-w9xqz"] Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.609585 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65fb5d464f-w9xqz"] Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.895156 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.926166 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d77c9\" (UniqueName: \"kubernetes.io/projected/45bb5bc6-b9e2-403f-9811-8223b2b660cb-kube-api-access-d77c9\") pod \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.926489 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-utilities\") pod \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.926567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-catalog-content\") pod \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\" (UID: \"45bb5bc6-b9e2-403f-9811-8223b2b660cb\") " Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.927556 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-utilities" (OuterVolumeSpecName: "utilities") pod "45bb5bc6-b9e2-403f-9811-8223b2b660cb" (UID: "45bb5bc6-b9e2-403f-9811-8223b2b660cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:35:19 crc kubenswrapper[4758]: I1203 18:35:19.930140 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45bb5bc6-b9e2-403f-9811-8223b2b660cb-kube-api-access-d77c9" (OuterVolumeSpecName: "kube-api-access-d77c9") pod "45bb5bc6-b9e2-403f-9811-8223b2b660cb" (UID: "45bb5bc6-b9e2-403f-9811-8223b2b660cb"). InnerVolumeSpecName "kube-api-access-d77c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.029269 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d77c9\" (UniqueName: \"kubernetes.io/projected/45bb5bc6-b9e2-403f-9811-8223b2b660cb-kube-api-access-d77c9\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.029308 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.058338 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45bb5bc6-b9e2-403f-9811-8223b2b660cb" (UID: "45bb5bc6-b9e2-403f-9811-8223b2b660cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.114836 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:35:20 crc kubenswrapper[4758]: E1203 18:35:20.115196 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.132858 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45bb5bc6-b9e2-403f-9811-8223b2b660cb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.150597 4758 generic.go:334] "Generic (PLEG): container finished" podID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerID="7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb" exitCode=0 Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.150691 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8ltr" event={"ID":"45bb5bc6-b9e2-403f-9811-8223b2b660cb","Type":"ContainerDied","Data":"7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb"} Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.150706 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8ltr" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.150737 4758 scope.go:117] "RemoveContainer" containerID="7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.150725 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8ltr" event={"ID":"45bb5bc6-b9e2-403f-9811-8223b2b660cb","Type":"ContainerDied","Data":"46e73544c10d2de2db2df61b8f6efd3fe6688b06843de1e3349a2bce467aab23"} Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.182775 4758 scope.go:117] "RemoveContainer" containerID="5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.183412 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m8ltr"] Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.193649 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m8ltr"] Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.208744 4758 scope.go:117] "RemoveContainer" containerID="8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.229765 4758 scope.go:117] "RemoveContainer" containerID="7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb" Dec 03 18:35:20 crc kubenswrapper[4758]: E1203 18:35:20.230167 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb\": container with ID starting with 7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb not found: ID does not exist" containerID="7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.230205 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb"} err="failed to get container status \"7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb\": rpc error: code = NotFound desc = could not find container \"7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb\": container with ID starting with 7374bc9a64d7cf26fa2e61bb9e8e430bdfe9d8c3045296dfb797af651a852cbb not found: ID does not exist" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.230231 4758 scope.go:117] "RemoveContainer" containerID="5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1" Dec 03 18:35:20 crc kubenswrapper[4758]: E1203 18:35:20.230494 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1\": container with ID starting with 5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1 not found: ID does not exist" containerID="5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.230539 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1"} err="failed to get container status \"5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1\": rpc error: code = NotFound desc = could not find container \"5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1\": container with ID starting with 5c8fc448ed7db5c122906c0f4308e31d0f80310f2cfc9c9a1afaf097a40254d1 not found: ID does not exist" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.230560 4758 scope.go:117] "RemoveContainer" containerID="8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3" Dec 03 18:35:20 crc kubenswrapper[4758]: E1203 18:35:20.230847 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3\": container with ID starting with 8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3 not found: ID does not exist" containerID="8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.230876 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3"} err="failed to get container status \"8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3\": rpc error: code = NotFound desc = could not find container \"8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3\": container with ID starting with 8f3faa2a03ddff0b66f2c47937374153c8aa3ab0084af0003e769ee5a5c3a8a3 not found: ID does not exist" Dec 03 18:35:20 crc kubenswrapper[4758]: I1203 18:35:20.910032 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.127623 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" path="/var/lib/kubelet/pods/21bc458f-4b56-4ed0-b3e4-fbad588d8669/volumes" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.128570 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" path="/var/lib/kubelet/pods/45bb5bc6-b9e2-403f-9811-8223b2b660cb/volumes" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.341978 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.342046 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.594407 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.621712 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.816412 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.833851 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.833903 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 18:35:21 crc kubenswrapper[4758]: I1203 18:35:21.835491 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:22 crc kubenswrapper[4758]: I1203 18:35:22.199601 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 18:35:22 crc kubenswrapper[4758]: I1203 18:35:22.250669 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 18:35:22 crc kubenswrapper[4758]: I1203 18:35:22.428936 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:35:22 crc kubenswrapper[4758]: I1203 18:35:22.428943 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:35:22 crc kubenswrapper[4758]: I1203 18:35:22.917887 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.88:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:35:22 crc kubenswrapper[4758]: I1203 18:35:22.917938 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.88:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:35:23 crc kubenswrapper[4758]: I1203 18:35:23.541232 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 18:35:27 crc kubenswrapper[4758]: I1203 18:35:27.509952 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.114411 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:35:31 crc kubenswrapper[4758]: E1203 18:35:31.115143 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.606593 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.611431 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.613363 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.837594 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.839012 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.839197 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 18:35:31 crc kubenswrapper[4758]: I1203 18:35:31.841642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 18:35:32 crc kubenswrapper[4758]: I1203 18:35:32.624166 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 18:35:32 crc kubenswrapper[4758]: I1203 18:35:32.626555 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 18:35:32 crc kubenswrapper[4758]: I1203 18:35:32.628078 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.163832 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:40 crc kubenswrapper[4758]: E1203 18:35:40.166083 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="registry-server" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.166199 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="registry-server" Dec 03 18:35:40 crc kubenswrapper[4758]: E1203 18:35:40.166297 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="extract-content" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.166369 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="extract-content" Dec 03 18:35:40 crc kubenswrapper[4758]: E1203 18:35:40.166445 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerName="init" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.166525 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerName="init" Dec 03 18:35:40 crc kubenswrapper[4758]: E1203 18:35:40.166604 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerName="dnsmasq-dns" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.166704 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerName="dnsmasq-dns" Dec 03 18:35:40 crc kubenswrapper[4758]: E1203 18:35:40.166818 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="extract-utilities" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.166901 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="extract-utilities" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.167214 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="21bc458f-4b56-4ed0-b3e4-fbad588d8669" containerName="dnsmasq-dns" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.167329 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="45bb5bc6-b9e2-403f-9811-8223b2b660cb" containerName="registry-server" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.168592 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.172022 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.176042 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.240975 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-scripts\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.241030 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01a20623-b5a5-404f-9a1d-ddab6f52f610-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.241059 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.241095 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s4lc\" (UniqueName: \"kubernetes.io/projected/01a20623-b5a5-404f-9a1d-ddab6f52f610-kube-api-access-9s4lc\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.241172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.241248 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342405 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-scripts\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01a20623-b5a5-404f-9a1d-ddab6f52f610-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342481 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342512 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s4lc\" (UniqueName: \"kubernetes.io/projected/01a20623-b5a5-404f-9a1d-ddab6f52f610-kube-api-access-9s4lc\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342612 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.342762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01a20623-b5a5-404f-9a1d-ddab6f52f610-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.348761 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-scripts\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.348919 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.349067 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.349218 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.358821 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s4lc\" (UniqueName: \"kubernetes.io/projected/01a20623-b5a5-404f-9a1d-ddab6f52f610-kube-api-access-9s4lc\") pod \"cinder-scheduler-0\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.508978 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 18:35:40 crc kubenswrapper[4758]: I1203 18:35:40.967977 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:41 crc kubenswrapper[4758]: I1203 18:35:41.697887 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:41 crc kubenswrapper[4758]: I1203 18:35:41.698448 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api-log" containerID="cri-o://ee58edb46c1472722256f05d635bc1ee927748b55b15a59dd645861e18a9ed0a" gracePeriod=30 Dec 03 18:35:41 crc kubenswrapper[4758]: I1203 18:35:41.698563 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api" containerID="cri-o://a9e5ebad6b7d6ba9e1dc792a00debe6467eedb6089ee5c19f2a948d6613d6e90" gracePeriod=30 Dec 03 18:35:41 crc kubenswrapper[4758]: I1203 18:35:41.720879 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01a20623-b5a5-404f-9a1d-ddab6f52f610","Type":"ContainerStarted","Data":"54116e4961c4b4a7a767d9c8e6f9ef4cee70902145004b70ae7ca134093c34ec"} Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.114087 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:35:42 crc kubenswrapper[4758]: E1203 18:35:42.114644 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.329905 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.331305 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.337406 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.368080 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490618 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490689 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490717 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490743 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-run\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490768 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8hxq\" (UniqueName: \"kubernetes.io/projected/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-kube-api-access-v8hxq\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490810 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490852 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490950 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490970 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.490986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.491003 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.491022 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.491131 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.491191 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.491217 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-run\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593281 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593318 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8hxq\" (UniqueName: \"kubernetes.io/projected/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-kube-api-access-v8hxq\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593344 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-run\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593348 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593726 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.593599 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595407 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595431 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595496 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595554 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595556 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595605 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595726 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595734 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595808 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.595921 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596031 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596070 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596105 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596096 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596126 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596172 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.596179 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.601256 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.601645 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.601777 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.607191 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.607470 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.612922 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8hxq\" (UniqueName: \"kubernetes.io/projected/779faeb6-b6e1-4953-9c93-41fc7e90dcb7-kube-api-access-v8hxq\") pod \"cinder-volume-volume1-0\" (UID: \"779faeb6-b6e1-4953-9c93-41fc7e90dcb7\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.652977 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.743712 4758 generic.go:334] "Generic (PLEG): container finished" podID="c942e361-c212-420a-903a-e6977f8fb011" containerID="ee58edb46c1472722256f05d635bc1ee927748b55b15a59dd645861e18a9ed0a" exitCode=143 Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.743798 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c942e361-c212-420a-903a-e6977f8fb011","Type":"ContainerDied","Data":"ee58edb46c1472722256f05d635bc1ee927748b55b15a59dd645861e18a9ed0a"} Dec 03 18:35:42 crc kubenswrapper[4758]: I1203 18:35:42.747690 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01a20623-b5a5-404f-9a1d-ddab6f52f610","Type":"ContainerStarted","Data":"9c249dbb8c03e8ba065ccdfc102a0338fc1df803b174ff377fb8e8744f23f8e3"} Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.016812 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.018916 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.020933 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.033018 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112097 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-dev\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112181 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-sys\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112219 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112288 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112314 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112341 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112442 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112488 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-scripts\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112504 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112525 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-lib-modules\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112554 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-run\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112602 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-ceph\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112636 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2bq5\" (UniqueName: \"kubernetes.io/projected/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-kube-api-access-d2bq5\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112702 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.112726 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-config-data\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214175 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-sys\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214221 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214271 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214289 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214329 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214334 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-sys\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214354 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214404 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214462 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214506 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-scripts\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214581 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214611 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214741 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214773 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-lib-modules\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-run\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214782 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214876 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-ceph\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214885 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-lib-modules\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.214948 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-run\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.215565 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2bq5\" (UniqueName: \"kubernetes.io/projected/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-kube-api-access-d2bq5\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.215636 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.215661 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-config-data\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.215795 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-dev\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.215830 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-dev\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.215870 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.222355 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-scripts\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.222373 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.222799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-ceph\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.223094 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-config-data\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.225531 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.235026 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2bq5\" (UniqueName: \"kubernetes.io/projected/e783b83e-d408-4f5b-b2ba-c322f8bd93f5-kube-api-access-d2bq5\") pod \"cinder-backup-0\" (UID: \"e783b83e-d408-4f5b-b2ba-c322f8bd93f5\") " pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: W1203 18:35:43.295447 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779faeb6_b6e1_4953_9c93_41fc7e90dcb7.slice/crio-af6bbd7fc8c0caf03a8642bc18ece23f6917c4b4c025d326ebd626dd41ddb45b WatchSource:0}: Error finding container af6bbd7fc8c0caf03a8642bc18ece23f6917c4b4c025d326ebd626dd41ddb45b: Status 404 returned error can't find the container with id af6bbd7fc8c0caf03a8642bc18ece23f6917c4b4c025d326ebd626dd41ddb45b Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.296059 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.357776 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.762128 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01a20623-b5a5-404f-9a1d-ddab6f52f610","Type":"ContainerStarted","Data":"b99756c905e08e88120a91e536300bccdf4ac414a785188144e47d89c9200e7c"} Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.764345 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"779faeb6-b6e1-4953-9c93-41fc7e90dcb7","Type":"ContainerStarted","Data":"af6bbd7fc8c0caf03a8642bc18ece23f6917c4b4c025d326ebd626dd41ddb45b"} Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.803697 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.803658677 podStartE2EDuration="3.803658677s" podCreationTimestamp="2025-12-03 18:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:43.796171345 +0000 UTC m=+5998.997548206" watchObservedRunningTime="2025-12-03 18:35:43.803658677 +0000 UTC m=+5999.005035538" Dec 03 18:35:43 crc kubenswrapper[4758]: I1203 18:35:43.898887 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 18:35:44 crc kubenswrapper[4758]: I1203 18:35:44.776371 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"e783b83e-d408-4f5b-b2ba-c322f8bd93f5","Type":"ContainerStarted","Data":"f9b5055ec601c1771bbcd16097533ea6ac77980c2ef36fc4723746c9c2b12f3f"} Dec 03 18:35:44 crc kubenswrapper[4758]: I1203 18:35:44.819384 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.84:8776/healthcheck\": read tcp 10.217.0.2:43034->10.217.1.84:8776: read: connection reset by peer" Dec 03 18:35:45 crc kubenswrapper[4758]: I1203 18:35:45.510035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 18:35:45 crc kubenswrapper[4758]: I1203 18:35:45.788383 4758 generic.go:334] "Generic (PLEG): container finished" podID="c942e361-c212-420a-903a-e6977f8fb011" containerID="a9e5ebad6b7d6ba9e1dc792a00debe6467eedb6089ee5c19f2a948d6613d6e90" exitCode=0 Dec 03 18:35:45 crc kubenswrapper[4758]: I1203 18:35:45.788432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c942e361-c212-420a-903a-e6977f8fb011","Type":"ContainerDied","Data":"a9e5ebad6b7d6ba9e1dc792a00debe6467eedb6089ee5c19f2a948d6613d6e90"} Dec 03 18:35:46 crc kubenswrapper[4758]: I1203 18:35:46.984093 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.089881 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c942e361-c212-420a-903a-e6977f8fb011-etc-machine-id\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.089992 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c942e361-c212-420a-903a-e6977f8fb011-logs\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.090057 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-combined-ca-bundle\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.090098 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.090143 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ttf2\" (UniqueName: \"kubernetes.io/projected/c942e361-c212-420a-903a-e6977f8fb011-kube-api-access-7ttf2\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.090311 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data-custom\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.090306 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c942e361-c212-420a-903a-e6977f8fb011-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.090373 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-scripts\") pod \"c942e361-c212-420a-903a-e6977f8fb011\" (UID: \"c942e361-c212-420a-903a-e6977f8fb011\") " Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.091003 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c942e361-c212-420a-903a-e6977f8fb011-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.091421 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c942e361-c212-420a-903a-e6977f8fb011-logs" (OuterVolumeSpecName: "logs") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.104905 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.104924 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-scripts" (OuterVolumeSpecName: "scripts") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.105021 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c942e361-c212-420a-903a-e6977f8fb011-kube-api-access-7ttf2" (OuterVolumeSpecName: "kube-api-access-7ttf2") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "kube-api-access-7ttf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.130858 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.160394 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data" (OuterVolumeSpecName: "config-data") pod "c942e361-c212-420a-903a-e6977f8fb011" (UID: "c942e361-c212-420a-903a-e6977f8fb011"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.193315 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.193358 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c942e361-c212-420a-903a-e6977f8fb011-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.193372 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.193384 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.193394 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ttf2\" (UniqueName: \"kubernetes.io/projected/c942e361-c212-420a-903a-e6977f8fb011-kube-api-access-7ttf2\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.193458 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c942e361-c212-420a-903a-e6977f8fb011-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.810443 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c942e361-c212-420a-903a-e6977f8fb011","Type":"ContainerDied","Data":"049e0a7dde4e8f240e2fda13a73f1f2c5e7119244db1c6acc4bd0a968aef080d"} Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.810493 4758 scope.go:117] "RemoveContainer" containerID="a9e5ebad6b7d6ba9e1dc792a00debe6467eedb6089ee5c19f2a948d6613d6e90" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.810516 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.839106 4758 scope.go:117] "RemoveContainer" containerID="ee58edb46c1472722256f05d635bc1ee927748b55b15a59dd645861e18a9ed0a" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.843426 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.855368 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.867025 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:47 crc kubenswrapper[4758]: E1203 18:35:47.867454 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api-log" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.867469 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api-log" Dec 03 18:35:47 crc kubenswrapper[4758]: E1203 18:35:47.867543 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.867550 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.867777 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api-log" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.867792 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c942e361-c212-420a-903a-e6977f8fb011" containerName="cinder-api" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.869350 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.872141 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 18:35:47 crc kubenswrapper[4758]: I1203 18:35:47.887388 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.009784 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10679416-8851-4d9c-9450-6722e89c06ba-logs\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.009844 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-scripts\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.009874 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbp4\" (UniqueName: \"kubernetes.io/projected/10679416-8851-4d9c-9450-6722e89c06ba-kube-api-access-sgbp4\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.010182 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.010646 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.010891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10679416-8851-4d9c-9450-6722e89c06ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.011059 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-config-data\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.112911 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113014 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10679416-8851-4d9c-9450-6722e89c06ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113061 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-config-data\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113096 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10679416-8851-4d9c-9450-6722e89c06ba-logs\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113161 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-scripts\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113170 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10679416-8851-4d9c-9450-6722e89c06ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113198 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbp4\" (UniqueName: \"kubernetes.io/projected/10679416-8851-4d9c-9450-6722e89c06ba-kube-api-access-sgbp4\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.113537 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.114077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10679416-8851-4d9c-9450-6722e89c06ba-logs\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.127460 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.127920 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-scripts\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.127947 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.128656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10679416-8851-4d9c-9450-6722e89c06ba-config-data\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.144249 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbp4\" (UniqueName: \"kubernetes.io/projected/10679416-8851-4d9c-9450-6722e89c06ba-kube-api-access-sgbp4\") pod \"cinder-api-0\" (UID: \"10679416-8851-4d9c-9450-6722e89c06ba\") " pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.200756 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 18:35:48 crc kubenswrapper[4758]: W1203 18:35:48.728976 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10679416_8851_4d9c_9450_6722e89c06ba.slice/crio-c3302536cb4fb5690bc906d3016eab0abbc2bffda881cd905845ece2ad38e0a0 WatchSource:0}: Error finding container c3302536cb4fb5690bc906d3016eab0abbc2bffda881cd905845ece2ad38e0a0: Status 404 returned error can't find the container with id c3302536cb4fb5690bc906d3016eab0abbc2bffda881cd905845ece2ad38e0a0 Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.734284 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 18:35:48 crc kubenswrapper[4758]: I1203 18:35:48.823217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10679416-8851-4d9c-9450-6722e89c06ba","Type":"ContainerStarted","Data":"c3302536cb4fb5690bc906d3016eab0abbc2bffda881cd905845ece2ad38e0a0"} Dec 03 18:35:49 crc kubenswrapper[4758]: I1203 18:35:49.130411 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c942e361-c212-420a-903a-e6977f8fb011" path="/var/lib/kubelet/pods/c942e361-c212-420a-903a-e6977f8fb011/volumes" Dec 03 18:35:49 crc kubenswrapper[4758]: I1203 18:35:49.837929 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"779faeb6-b6e1-4953-9c93-41fc7e90dcb7","Type":"ContainerStarted","Data":"f730cb5fdd430c6d101ab9c9023c40717618ebbe3f4578cafd25124fa2533084"} Dec 03 18:35:50 crc kubenswrapper[4758]: I1203 18:35:50.698224 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 18:35:50 crc kubenswrapper[4758]: I1203 18:35:50.763077 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:50 crc kubenswrapper[4758]: I1203 18:35:50.847220 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="cinder-scheduler" containerID="cri-o://9c249dbb8c03e8ba065ccdfc102a0338fc1df803b174ff377fb8e8744f23f8e3" gracePeriod=30 Dec 03 18:35:50 crc kubenswrapper[4758]: I1203 18:35:50.847313 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="probe" containerID="cri-o://b99756c905e08e88120a91e536300bccdf4ac414a785188144e47d89c9200e7c" gracePeriod=30 Dec 03 18:35:51 crc kubenswrapper[4758]: I1203 18:35:51.857887 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"779faeb6-b6e1-4953-9c93-41fc7e90dcb7","Type":"ContainerStarted","Data":"bb7490d31137ba749847938ee969fd2009db898d9d3d22932c66c41aa772ae11"} Dec 03 18:35:51 crc kubenswrapper[4758]: I1203 18:35:51.861072 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10679416-8851-4d9c-9450-6722e89c06ba","Type":"ContainerStarted","Data":"1c5f59269806d13686a6528117ad4c79a0ad524ed6a022e31024cb8156bd7f0a"} Dec 03 18:35:51 crc kubenswrapper[4758]: I1203 18:35:51.905655 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.629587681 podStartE2EDuration="9.905623621s" podCreationTimestamp="2025-12-03 18:35:42 +0000 UTC" firstStartedPulling="2025-12-03 18:35:43.297585141 +0000 UTC m=+5998.498962002" lastFinishedPulling="2025-12-03 18:35:48.573621081 +0000 UTC m=+6003.774997942" observedRunningTime="2025-12-03 18:35:51.898760346 +0000 UTC m=+6007.100137207" watchObservedRunningTime="2025-12-03 18:35:51.905623621 +0000 UTC m=+6007.107000482" Dec 03 18:35:52 crc kubenswrapper[4758]: I1203 18:35:52.654790 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:52 crc kubenswrapper[4758]: I1203 18:35:52.874494 4758 generic.go:334] "Generic (PLEG): container finished" podID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerID="b99756c905e08e88120a91e536300bccdf4ac414a785188144e47d89c9200e7c" exitCode=0 Dec 03 18:35:52 crc kubenswrapper[4758]: I1203 18:35:52.874535 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01a20623-b5a5-404f-9a1d-ddab6f52f610","Type":"ContainerDied","Data":"b99756c905e08e88120a91e536300bccdf4ac414a785188144e47d89c9200e7c"} Dec 03 18:35:53 crc kubenswrapper[4758]: I1203 18:35:53.884791 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"e783b83e-d408-4f5b-b2ba-c322f8bd93f5","Type":"ContainerStarted","Data":"e77278190328a674c9ef396aef41c2951eb18838533b419e256c3e75e5eb2a9d"} Dec 03 18:35:53 crc kubenswrapper[4758]: I1203 18:35:53.886949 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10679416-8851-4d9c-9450-6722e89c06ba","Type":"ContainerStarted","Data":"1e3b0a0824c960f107757f8b657737885fb80fef5a2350534b917ea375f836f9"} Dec 03 18:35:55 crc kubenswrapper[4758]: I1203 18:35:55.920450 4758 generic.go:334] "Generic (PLEG): container finished" podID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerID="9c249dbb8c03e8ba065ccdfc102a0338fc1df803b174ff377fb8e8744f23f8e3" exitCode=0 Dec 03 18:35:55 crc kubenswrapper[4758]: I1203 18:35:55.920635 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01a20623-b5a5-404f-9a1d-ddab6f52f610","Type":"ContainerDied","Data":"9c249dbb8c03e8ba065ccdfc102a0338fc1df803b174ff377fb8e8744f23f8e3"} Dec 03 18:35:55 crc kubenswrapper[4758]: I1203 18:35:55.921135 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 18:35:55 crc kubenswrapper[4758]: I1203 18:35:55.950801 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.950777406 podStartE2EDuration="8.950777406s" podCreationTimestamp="2025-12-03 18:35:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:35:55.942483892 +0000 UTC m=+6011.143860753" watchObservedRunningTime="2025-12-03 18:35:55.950777406 +0000 UTC m=+6011.152154257" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.545703 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.583404 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-scripts\") pod \"01a20623-b5a5-404f-9a1d-ddab6f52f610\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.583472 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-combined-ca-bundle\") pod \"01a20623-b5a5-404f-9a1d-ddab6f52f610\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.583608 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data\") pod \"01a20623-b5a5-404f-9a1d-ddab6f52f610\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.583766 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01a20623-b5a5-404f-9a1d-ddab6f52f610-etc-machine-id\") pod \"01a20623-b5a5-404f-9a1d-ddab6f52f610\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.583807 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s4lc\" (UniqueName: \"kubernetes.io/projected/01a20623-b5a5-404f-9a1d-ddab6f52f610-kube-api-access-9s4lc\") pod \"01a20623-b5a5-404f-9a1d-ddab6f52f610\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.583838 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data-custom\") pod \"01a20623-b5a5-404f-9a1d-ddab6f52f610\" (UID: \"01a20623-b5a5-404f-9a1d-ddab6f52f610\") " Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.584885 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01a20623-b5a5-404f-9a1d-ddab6f52f610-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "01a20623-b5a5-404f-9a1d-ddab6f52f610" (UID: "01a20623-b5a5-404f-9a1d-ddab6f52f610"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.589854 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-scripts" (OuterVolumeSpecName: "scripts") pod "01a20623-b5a5-404f-9a1d-ddab6f52f610" (UID: "01a20623-b5a5-404f-9a1d-ddab6f52f610"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.590163 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "01a20623-b5a5-404f-9a1d-ddab6f52f610" (UID: "01a20623-b5a5-404f-9a1d-ddab6f52f610"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.590178 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a20623-b5a5-404f-9a1d-ddab6f52f610-kube-api-access-9s4lc" (OuterVolumeSpecName: "kube-api-access-9s4lc") pod "01a20623-b5a5-404f-9a1d-ddab6f52f610" (UID: "01a20623-b5a5-404f-9a1d-ddab6f52f610"). InnerVolumeSpecName "kube-api-access-9s4lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.658120 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01a20623-b5a5-404f-9a1d-ddab6f52f610" (UID: "01a20623-b5a5-404f-9a1d-ddab6f52f610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.677914 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data" (OuterVolumeSpecName: "config-data") pod "01a20623-b5a5-404f-9a1d-ddab6f52f610" (UID: "01a20623-b5a5-404f-9a1d-ddab6f52f610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.686152 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.686186 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01a20623-b5a5-404f-9a1d-ddab6f52f610-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.686198 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s4lc\" (UniqueName: \"kubernetes.io/projected/01a20623-b5a5-404f-9a1d-ddab6f52f610-kube-api-access-9s4lc\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.686343 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.686359 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.686369 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a20623-b5a5-404f-9a1d-ddab6f52f610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.931568 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01a20623-b5a5-404f-9a1d-ddab6f52f610","Type":"ContainerDied","Data":"54116e4961c4b4a7a767d9c8e6f9ef4cee70902145004b70ae7ca134093c34ec"} Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.931629 4758 scope.go:117] "RemoveContainer" containerID="b99756c905e08e88120a91e536300bccdf4ac414a785188144e47d89c9200e7c" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.931661 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.935052 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"e783b83e-d408-4f5b-b2ba-c322f8bd93f5","Type":"ContainerStarted","Data":"a3b5a3778904c27536a692d2d55daf52650b1818ea50c1eef72ec7e44d02cc73"} Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.983206 4758 scope.go:117] "RemoveContainer" containerID="9c249dbb8c03e8ba065ccdfc102a0338fc1df803b174ff377fb8e8744f23f8e3" Dec 03 18:35:56 crc kubenswrapper[4758]: I1203 18:35:56.993964 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=6.344151537 podStartE2EDuration="14.993930508s" podCreationTimestamp="2025-12-03 18:35:42 +0000 UTC" firstStartedPulling="2025-12-03 18:35:43.910298415 +0000 UTC m=+5999.111675276" lastFinishedPulling="2025-12-03 18:35:52.560077386 +0000 UTC m=+6007.761454247" observedRunningTime="2025-12-03 18:35:56.975597581 +0000 UTC m=+6012.176974462" watchObservedRunningTime="2025-12-03 18:35:56.993930508 +0000 UTC m=+6012.195307379" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.018781 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.034150 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.043094 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:57 crc kubenswrapper[4758]: E1203 18:35:57.043966 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="cinder-scheduler" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.043990 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="cinder-scheduler" Dec 03 18:35:57 crc kubenswrapper[4758]: E1203 18:35:57.044040 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="probe" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.044051 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="probe" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.044265 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="cinder-scheduler" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.044305 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" containerName="probe" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.046057 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.051094 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.052836 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.099164 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.099221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.099362 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.099468 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfc5m\" (UniqueName: \"kubernetes.io/projected/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-kube-api-access-mfc5m\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.099606 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.099908 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.114561 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:35:57 crc kubenswrapper[4758]: E1203 18:35:57.114772 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.126702 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a20623-b5a5-404f-9a1d-ddab6f52f610" path="/var/lib/kubelet/pods/01a20623-b5a5-404f-9a1d-ddab6f52f610/volumes" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.201527 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.201698 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.201740 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.201775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.201806 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfc5m\" (UniqueName: \"kubernetes.io/projected/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-kube-api-access-mfc5m\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.201852 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.203045 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.206668 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.206866 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.207565 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.208539 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.222765 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfc5m\" (UniqueName: \"kubernetes.io/projected/8d01ab7c-bf47-4378-b41c-508c5e71e0b4-kube-api-access-mfc5m\") pod \"cinder-scheduler-0\" (UID: \"8d01ab7c-bf47-4378-b41c-508c5e71e0b4\") " pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.368625 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.844400 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.853111 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 03 18:35:57 crc kubenswrapper[4758]: I1203 18:35:57.950841 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8d01ab7c-bf47-4378-b41c-508c5e71e0b4","Type":"ContainerStarted","Data":"5d02e3cc80eb82bf994f03c7ebcd10c1d6a2e59fdcfb0d3fa9270e6f06ff080e"} Dec 03 18:35:58 crc kubenswrapper[4758]: I1203 18:35:58.357941 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 03 18:35:58 crc kubenswrapper[4758]: I1203 18:35:58.571180 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 03 18:35:59 crc kubenswrapper[4758]: I1203 18:35:59.970053 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8d01ab7c-bf47-4378-b41c-508c5e71e0b4","Type":"ContainerStarted","Data":"1e9892edf9488a15022902023fca2dbac21cacce030e8281abb3f5a4d0a473c8"} Dec 03 18:36:00 crc kubenswrapper[4758]: I1203 18:36:00.260986 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 18:36:03 crc kubenswrapper[4758]: I1203 18:36:03.001261 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8d01ab7c-bf47-4378-b41c-508c5e71e0b4","Type":"ContainerStarted","Data":"5ec9e5c7ad0907184ab98372eb81016978e362e58c060238ad60db973aaa1bc3"} Dec 03 18:36:03 crc kubenswrapper[4758]: I1203 18:36:03.027271 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.027254647 podStartE2EDuration="6.027254647s" podCreationTimestamp="2025-12-03 18:35:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:36:03.018257363 +0000 UTC m=+6018.219634234" watchObservedRunningTime="2025-12-03 18:36:03.027254647 +0000 UTC m=+6018.228631508" Dec 03 18:36:07 crc kubenswrapper[4758]: I1203 18:36:07.370035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 18:36:07 crc kubenswrapper[4758]: I1203 18:36:07.563485 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 18:36:08 crc kubenswrapper[4758]: I1203 18:36:08.115801 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:36:08 crc kubenswrapper[4758]: E1203 18:36:08.116040 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:36:19 crc kubenswrapper[4758]: I1203 18:36:19.114316 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:36:19 crc kubenswrapper[4758]: E1203 18:36:19.115217 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:36:30 crc kubenswrapper[4758]: I1203 18:36:30.115084 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:36:30 crc kubenswrapper[4758]: E1203 18:36:30.115986 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:36:45 crc kubenswrapper[4758]: I1203 18:36:45.128234 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:36:45 crc kubenswrapper[4758]: E1203 18:36:45.129370 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:36:50 crc kubenswrapper[4758]: I1203 18:36:50.923535 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-457qh"] Dec 03 18:36:50 crc kubenswrapper[4758]: I1203 18:36:50.931866 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:50 crc kubenswrapper[4758]: I1203 18:36:50.999362 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-457qh"] Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.024141 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsgwn\" (UniqueName: \"kubernetes.io/projected/0cea688c-c7db-47ff-b1c9-28d5795fbb65-kube-api-access-zsgwn\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.024204 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-catalog-content\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.024290 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-utilities\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.126305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsgwn\" (UniqueName: \"kubernetes.io/projected/0cea688c-c7db-47ff-b1c9-28d5795fbb65-kube-api-access-zsgwn\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.126357 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-catalog-content\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.126404 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-utilities\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.127090 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-utilities\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.127285 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-catalog-content\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.143725 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsgwn\" (UniqueName: \"kubernetes.io/projected/0cea688c-c7db-47ff-b1c9-28d5795fbb65-kube-api-access-zsgwn\") pod \"community-operators-457qh\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.308640 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:36:51 crc kubenswrapper[4758]: I1203 18:36:51.749519 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-457qh"] Dec 03 18:36:52 crc kubenswrapper[4758]: I1203 18:36:52.529710 4758 generic.go:334] "Generic (PLEG): container finished" podID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerID="bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2" exitCode=0 Dec 03 18:36:52 crc kubenswrapper[4758]: I1203 18:36:52.529771 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerDied","Data":"bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2"} Dec 03 18:36:52 crc kubenswrapper[4758]: I1203 18:36:52.530005 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerStarted","Data":"00b53b5a4418db884058dd76aacdeaf930591e9dbedf3c21413c7b12f21a628f"} Dec 03 18:36:56 crc kubenswrapper[4758]: I1203 18:36:56.584041 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerStarted","Data":"0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02"} Dec 03 18:36:57 crc kubenswrapper[4758]: I1203 18:36:57.594870 4758 generic.go:334] "Generic (PLEG): container finished" podID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerID="0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02" exitCode=0 Dec 03 18:36:57 crc kubenswrapper[4758]: I1203 18:36:57.594938 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerDied","Data":"0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02"} Dec 03 18:36:59 crc kubenswrapper[4758]: I1203 18:36:59.118613 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:36:59 crc kubenswrapper[4758]: E1203 18:36:59.119248 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:36:59 crc kubenswrapper[4758]: I1203 18:36:59.624736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerStarted","Data":"ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988"} Dec 03 18:36:59 crc kubenswrapper[4758]: I1203 18:36:59.651047 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-457qh" podStartSLOduration=3.408078719 podStartE2EDuration="9.651030646s" podCreationTimestamp="2025-12-03 18:36:50 +0000 UTC" firstStartedPulling="2025-12-03 18:36:52.533161764 +0000 UTC m=+6067.734538635" lastFinishedPulling="2025-12-03 18:36:58.776113691 +0000 UTC m=+6073.977490562" observedRunningTime="2025-12-03 18:36:59.64524131 +0000 UTC m=+6074.846618171" watchObservedRunningTime="2025-12-03 18:36:59.651030646 +0000 UTC m=+6074.852407507" Dec 03 18:37:01 crc kubenswrapper[4758]: I1203 18:37:01.309608 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:37:01 crc kubenswrapper[4758]: I1203 18:37:01.309718 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:37:01 crc kubenswrapper[4758]: I1203 18:37:01.368449 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:37:11 crc kubenswrapper[4758]: I1203 18:37:11.369268 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:37:11 crc kubenswrapper[4758]: I1203 18:37:11.434282 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-457qh"] Dec 03 18:37:11 crc kubenswrapper[4758]: I1203 18:37:11.731833 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-457qh" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="registry-server" containerID="cri-o://ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988" gracePeriod=2 Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.216760 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.379438 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsgwn\" (UniqueName: \"kubernetes.io/projected/0cea688c-c7db-47ff-b1c9-28d5795fbb65-kube-api-access-zsgwn\") pod \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.379543 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-utilities\") pod \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.379705 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-catalog-content\") pod \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\" (UID: \"0cea688c-c7db-47ff-b1c9-28d5795fbb65\") " Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.381120 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-utilities" (OuterVolumeSpecName: "utilities") pod "0cea688c-c7db-47ff-b1c9-28d5795fbb65" (UID: "0cea688c-c7db-47ff-b1c9-28d5795fbb65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.386903 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cea688c-c7db-47ff-b1c9-28d5795fbb65-kube-api-access-zsgwn" (OuterVolumeSpecName: "kube-api-access-zsgwn") pod "0cea688c-c7db-47ff-b1c9-28d5795fbb65" (UID: "0cea688c-c7db-47ff-b1c9-28d5795fbb65"). InnerVolumeSpecName "kube-api-access-zsgwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.427359 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cea688c-c7db-47ff-b1c9-28d5795fbb65" (UID: "0cea688c-c7db-47ff-b1c9-28d5795fbb65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.481984 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.482016 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cea688c-c7db-47ff-b1c9-28d5795fbb65-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.482027 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsgwn\" (UniqueName: \"kubernetes.io/projected/0cea688c-c7db-47ff-b1c9-28d5795fbb65-kube-api-access-zsgwn\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.741965 4758 generic.go:334] "Generic (PLEG): container finished" podID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerID="ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988" exitCode=0 Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.742013 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerDied","Data":"ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988"} Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.742043 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-457qh" event={"ID":"0cea688c-c7db-47ff-b1c9-28d5795fbb65","Type":"ContainerDied","Data":"00b53b5a4418db884058dd76aacdeaf930591e9dbedf3c21413c7b12f21a628f"} Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.742062 4758 scope.go:117] "RemoveContainer" containerID="ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.742218 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-457qh" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.762961 4758 scope.go:117] "RemoveContainer" containerID="0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.781180 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-457qh"] Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.793568 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-457qh"] Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.804057 4758 scope.go:117] "RemoveContainer" containerID="bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.835869 4758 scope.go:117] "RemoveContainer" containerID="ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988" Dec 03 18:37:12 crc kubenswrapper[4758]: E1203 18:37:12.837101 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988\": container with ID starting with ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988 not found: ID does not exist" containerID="ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.837156 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988"} err="failed to get container status \"ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988\": rpc error: code = NotFound desc = could not find container \"ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988\": container with ID starting with ef2e96dc650f153ac80ec6187151535ea65651096ac7b79b7bf5afbce88d7988 not found: ID does not exist" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.837192 4758 scope.go:117] "RemoveContainer" containerID="0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02" Dec 03 18:37:12 crc kubenswrapper[4758]: E1203 18:37:12.837560 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02\": container with ID starting with 0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02 not found: ID does not exist" containerID="0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.837676 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02"} err="failed to get container status \"0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02\": rpc error: code = NotFound desc = could not find container \"0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02\": container with ID starting with 0cf595d8b2240c23fcab1c3536ffc947ef288465a3a5b20eb42569b520fd9b02 not found: ID does not exist" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.837792 4758 scope.go:117] "RemoveContainer" containerID="bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2" Dec 03 18:37:12 crc kubenswrapper[4758]: E1203 18:37:12.838162 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2\": container with ID starting with bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2 not found: ID does not exist" containerID="bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2" Dec 03 18:37:12 crc kubenswrapper[4758]: I1203 18:37:12.838195 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2"} err="failed to get container status \"bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2\": rpc error: code = NotFound desc = could not find container \"bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2\": container with ID starting with bfda255b82104f04d34af4efcc4d86b284f912a677dc45430fb910a4f28d35f2 not found: ID does not exist" Dec 03 18:37:13 crc kubenswrapper[4758]: I1203 18:37:13.075071 4758 scope.go:117] "RemoveContainer" containerID="c4557fe9f1c683b3bfb290f44a68ff42fcadadd41374bf6843b2ed7b87de8f5a" Dec 03 18:37:13 crc kubenswrapper[4758]: I1203 18:37:13.105880 4758 scope.go:117] "RemoveContainer" containerID="2c8d0cc9fc88bd1c2364a3dec00e634d67ffb87876839fbf99c771bc63915855" Dec 03 18:37:13 crc kubenswrapper[4758]: I1203 18:37:13.159993 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" path="/var/lib/kubelet/pods/0cea688c-c7db-47ff-b1c9-28d5795fbb65/volumes" Dec 03 18:37:14 crc kubenswrapper[4758]: I1203 18:37:14.053156 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jvjf4"] Dec 03 18:37:14 crc kubenswrapper[4758]: I1203 18:37:14.065914 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jvjf4"] Dec 03 18:37:14 crc kubenswrapper[4758]: I1203 18:37:14.078297 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8074-account-create-update-9cjh7"] Dec 03 18:37:14 crc kubenswrapper[4758]: I1203 18:37:14.090328 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8074-account-create-update-9cjh7"] Dec 03 18:37:14 crc kubenswrapper[4758]: I1203 18:37:14.114642 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:37:14 crc kubenswrapper[4758]: E1203 18:37:14.115130 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:37:15 crc kubenswrapper[4758]: I1203 18:37:15.125779 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2edbbc54-5da0-4483-bff6-6f8c919e9db2" path="/var/lib/kubelet/pods/2edbbc54-5da0-4483-bff6-6f8c919e9db2/volumes" Dec 03 18:37:15 crc kubenswrapper[4758]: I1203 18:37:15.126563 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53cff074-8895-4fd4-90f3-eebeea7c9caa" path="/var/lib/kubelet/pods/53cff074-8895-4fd4-90f3-eebeea7c9caa/volumes" Dec 03 18:37:21 crc kubenswrapper[4758]: I1203 18:37:21.028044 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-rpgg5"] Dec 03 18:37:21 crc kubenswrapper[4758]: I1203 18:37:21.037462 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-rpgg5"] Dec 03 18:37:21 crc kubenswrapper[4758]: I1203 18:37:21.131221 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="186e38e4-ec80-424f-bf8e-745158c88b21" path="/var/lib/kubelet/pods/186e38e4-ec80-424f-bf8e-745158c88b21/volumes" Dec 03 18:37:25 crc kubenswrapper[4758]: I1203 18:37:25.119443 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:37:25 crc kubenswrapper[4758]: E1203 18:37:25.119973 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:37:34 crc kubenswrapper[4758]: I1203 18:37:34.033323 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-spqsj"] Dec 03 18:37:34 crc kubenswrapper[4758]: I1203 18:37:34.045764 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-spqsj"] Dec 03 18:37:35 crc kubenswrapper[4758]: I1203 18:37:35.124421 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53831e41-c824-4d29-8d23-1c2db2e44ca9" path="/var/lib/kubelet/pods/53831e41-c824-4d29-8d23-1c2db2e44ca9/volumes" Dec 03 18:37:37 crc kubenswrapper[4758]: I1203 18:37:37.114425 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:37:37 crc kubenswrapper[4758]: E1203 18:37:37.115154 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:37:49 crc kubenswrapper[4758]: I1203 18:37:49.114791 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.176954 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"55064fb8d522abc0cbfb578b1ed4ff679982f3cf1154c3a5e9439b6fbc880f07"} Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.487704 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vts8k"] Dec 03 18:37:50 crc kubenswrapper[4758]: E1203 18:37:50.489124 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="extract-content" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.489244 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="extract-content" Dec 03 18:37:50 crc kubenswrapper[4758]: E1203 18:37:50.489323 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="registry-server" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.489375 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="registry-server" Dec 03 18:37:50 crc kubenswrapper[4758]: E1203 18:37:50.489470 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="extract-utilities" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.489542 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="extract-utilities" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.497317 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cea688c-c7db-47ff-b1c9-28d5795fbb65" containerName="registry-server" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.498493 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.506663 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jpntc" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.507333 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.520196 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-5jrck"] Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.522549 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.530960 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vts8k"] Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.542230 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5jrck"] Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.552273 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45509ece-d0bc-4fed-aaea-825badb17447-scripts\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.552323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-run-ovn\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.552366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-log-ovn\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.552412 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-run\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.552488 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pt6b\" (UniqueName: \"kubernetes.io/projected/45509ece-d0bc-4fed-aaea-825badb17447-kube-api-access-5pt6b\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655162 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-run\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655312 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-log\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655372 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-run\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655444 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-lib\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655490 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-etc-ovs\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655529 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pt6b\" (UniqueName: \"kubernetes.io/projected/45509ece-d0bc-4fed-aaea-825badb17447-kube-api-access-5pt6b\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655589 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45509ece-d0bc-4fed-aaea-825badb17447-scripts\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655635 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55wd8\" (UniqueName: \"kubernetes.io/projected/b829c7e2-ce7c-46f9-830b-aa7227537d2c-kube-api-access-55wd8\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655668 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-run-ovn\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655716 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b829c7e2-ce7c-46f9-830b-aa7227537d2c-scripts\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.655811 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-log-ovn\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.656378 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-log-ovn\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.656469 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-run\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.659172 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45509ece-d0bc-4fed-aaea-825badb17447-scripts\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.659284 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/45509ece-d0bc-4fed-aaea-825badb17447-var-run-ovn\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.676468 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pt6b\" (UniqueName: \"kubernetes.io/projected/45509ece-d0bc-4fed-aaea-825badb17447-kube-api-access-5pt6b\") pod \"ovn-controller-vts8k\" (UID: \"45509ece-d0bc-4fed-aaea-825badb17447\") " pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.757335 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-run\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.757776 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-etc-ovs\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.757803 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-lib\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.757874 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55wd8\" (UniqueName: \"kubernetes.io/projected/b829c7e2-ce7c-46f9-830b-aa7227537d2c-kube-api-access-55wd8\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.757900 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b829c7e2-ce7c-46f9-830b-aa7227537d2c-scripts\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.758036 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-log\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.758030 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-lib\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.758219 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-run\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.758229 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-var-log\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.760086 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b829c7e2-ce7c-46f9-830b-aa7227537d2c-scripts\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.760176 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b829c7e2-ce7c-46f9-830b-aa7227537d2c-etc-ovs\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.775379 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55wd8\" (UniqueName: \"kubernetes.io/projected/b829c7e2-ce7c-46f9-830b-aa7227537d2c-kube-api-access-55wd8\") pod \"ovn-controller-ovs-5jrck\" (UID: \"b829c7e2-ce7c-46f9-830b-aa7227537d2c\") " pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.833327 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vts8k" Dec 03 18:37:50 crc kubenswrapper[4758]: I1203 18:37:50.842008 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:51 crc kubenswrapper[4758]: I1203 18:37:51.364126 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vts8k"] Dec 03 18:37:51 crc kubenswrapper[4758]: W1203 18:37:51.381984 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45509ece_d0bc_4fed_aaea_825badb17447.slice/crio-49bf3844c87f341bd8eeb11657e6a14205fe37fc2f57924b6fd4890f26d227d5 WatchSource:0}: Error finding container 49bf3844c87f341bd8eeb11657e6a14205fe37fc2f57924b6fd4890f26d227d5: Status 404 returned error can't find the container with id 49bf3844c87f341bd8eeb11657e6a14205fe37fc2f57924b6fd4890f26d227d5 Dec 03 18:37:51 crc kubenswrapper[4758]: I1203 18:37:51.654026 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5jrck"] Dec 03 18:37:51 crc kubenswrapper[4758]: W1203 18:37:51.659287 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb829c7e2_ce7c_46f9_830b_aa7227537d2c.slice/crio-e5d14f9801ac9827cdbf85c65ae3247f8627c9cd983c58602d7eedc50a1fdef3 WatchSource:0}: Error finding container e5d14f9801ac9827cdbf85c65ae3247f8627c9cd983c58602d7eedc50a1fdef3: Status 404 returned error can't find the container with id e5d14f9801ac9827cdbf85c65ae3247f8627c9cd983c58602d7eedc50a1fdef3 Dec 03 18:37:51 crc kubenswrapper[4758]: I1203 18:37:51.966573 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-hzhjn"] Dec 03 18:37:51 crc kubenswrapper[4758]: I1203 18:37:51.968410 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:51 crc kubenswrapper[4758]: I1203 18:37:51.972558 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 18:37:51 crc kubenswrapper[4758]: I1203 18:37:51.983492 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hzhjn"] Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.084010 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws9gt\" (UniqueName: \"kubernetes.io/projected/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-kube-api-access-ws9gt\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.084070 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-ovs-rundir\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.084211 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-config\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.084279 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-ovn-rundir\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.185919 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws9gt\" (UniqueName: \"kubernetes.io/projected/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-kube-api-access-ws9gt\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.186267 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-ovs-rundir\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.186331 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-config\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.186431 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-ovn-rundir\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.186669 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-ovn-rundir\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.186666 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-ovs-rundir\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.187422 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-config\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.202245 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5jrck" event={"ID":"b829c7e2-ce7c-46f9-830b-aa7227537d2c","Type":"ContainerStarted","Data":"fd599916069c6a322ab9a9fe321848ae5c3721fe3c726c420bd51f0b0eafc777"} Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.202289 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5jrck" event={"ID":"b829c7e2-ce7c-46f9-830b-aa7227537d2c","Type":"ContainerStarted","Data":"e5d14f9801ac9827cdbf85c65ae3247f8627c9cd983c58602d7eedc50a1fdef3"} Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.204939 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws9gt\" (UniqueName: \"kubernetes.io/projected/a86f6acf-8cd6-41a9-a38a-1f113fe29d93-kube-api-access-ws9gt\") pod \"ovn-controller-metrics-hzhjn\" (UID: \"a86f6acf-8cd6-41a9-a38a-1f113fe29d93\") " pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.216855 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vts8k" event={"ID":"45509ece-d0bc-4fed-aaea-825badb17447","Type":"ContainerStarted","Data":"064e8007e5f8a717961fd3f8b81d0e1267840d348b942c9a9e0cc97f7f39ef77"} Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.216892 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vts8k" event={"ID":"45509ece-d0bc-4fed-aaea-825badb17447","Type":"ContainerStarted","Data":"49bf3844c87f341bd8eeb11657e6a14205fe37fc2f57924b6fd4890f26d227d5"} Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.217584 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vts8k" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.257954 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vts8k" podStartSLOduration=2.257930849 podStartE2EDuration="2.257930849s" podCreationTimestamp="2025-12-03 18:37:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:37:52.239653503 +0000 UTC m=+6127.441030364" watchObservedRunningTime="2025-12-03 18:37:52.257930849 +0000 UTC m=+6127.459307710" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.290338 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hzhjn" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.369013 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-lgqzq"] Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.370444 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.387540 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-lgqzq"] Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.492274 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh7rs\" (UniqueName: \"kubernetes.io/projected/412a365c-25b6-4058-9b1f-be4f76154b6a-kube-api-access-mh7rs\") pod \"octavia-db-create-lgqzq\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.492442 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412a365c-25b6-4058-9b1f-be4f76154b6a-operator-scripts\") pod \"octavia-db-create-lgqzq\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.594547 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh7rs\" (UniqueName: \"kubernetes.io/projected/412a365c-25b6-4058-9b1f-be4f76154b6a-kube-api-access-mh7rs\") pod \"octavia-db-create-lgqzq\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.594762 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412a365c-25b6-4058-9b1f-be4f76154b6a-operator-scripts\") pod \"octavia-db-create-lgqzq\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.595973 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412a365c-25b6-4058-9b1f-be4f76154b6a-operator-scripts\") pod \"octavia-db-create-lgqzq\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.611383 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh7rs\" (UniqueName: \"kubernetes.io/projected/412a365c-25b6-4058-9b1f-be4f76154b6a-kube-api-access-mh7rs\") pod \"octavia-db-create-lgqzq\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.743361 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hzhjn"] Dec 03 18:37:52 crc kubenswrapper[4758]: W1203 18:37:52.744253 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda86f6acf_8cd6_41a9_a38a_1f113fe29d93.slice/crio-a52f0cd09a63e5d6f8972b3fa50203c77392ab979ec49a15ba08b144c2540daa WatchSource:0}: Error finding container a52f0cd09a63e5d6f8972b3fa50203c77392ab979ec49a15ba08b144c2540daa: Status 404 returned error can't find the container with id a52f0cd09a63e5d6f8972b3fa50203c77392ab979ec49a15ba08b144c2540daa Dec 03 18:37:52 crc kubenswrapper[4758]: I1203 18:37:52.786793 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.238916 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hzhjn" event={"ID":"a86f6acf-8cd6-41a9-a38a-1f113fe29d93","Type":"ContainerStarted","Data":"016fb45610312bb9f683ec87c85698982c48cfb3f6c88f4f03bd47e329dc8874"} Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.239704 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hzhjn" event={"ID":"a86f6acf-8cd6-41a9-a38a-1f113fe29d93","Type":"ContainerStarted","Data":"a52f0cd09a63e5d6f8972b3fa50203c77392ab979ec49a15ba08b144c2540daa"} Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.242130 4758 generic.go:334] "Generic (PLEG): container finished" podID="b829c7e2-ce7c-46f9-830b-aa7227537d2c" containerID="fd599916069c6a322ab9a9fe321848ae5c3721fe3c726c420bd51f0b0eafc777" exitCode=0 Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.243734 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5jrck" event={"ID":"b829c7e2-ce7c-46f9-830b-aa7227537d2c","Type":"ContainerDied","Data":"fd599916069c6a322ab9a9fe321848ae5c3721fe3c726c420bd51f0b0eafc777"} Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.276789 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-lgqzq"] Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.315315 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-hzhjn" podStartSLOduration=2.315271365 podStartE2EDuration="2.315271365s" podCreationTimestamp="2025-12-03 18:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:37:53.263314118 +0000 UTC m=+6128.464690989" watchObservedRunningTime="2025-12-03 18:37:53.315271365 +0000 UTC m=+6128.516648236" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.669379 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-b699-account-create-update-j7gtf"] Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.671067 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.672682 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.679961 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-b699-account-create-update-j7gtf"] Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.723106 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4svl\" (UniqueName: \"kubernetes.io/projected/a25bce28-2142-4800-b3d6-e0059dae2e37-kube-api-access-p4svl\") pod \"octavia-b699-account-create-update-j7gtf\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.723179 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a25bce28-2142-4800-b3d6-e0059dae2e37-operator-scripts\") pod \"octavia-b699-account-create-update-j7gtf\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.824939 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4svl\" (UniqueName: \"kubernetes.io/projected/a25bce28-2142-4800-b3d6-e0059dae2e37-kube-api-access-p4svl\") pod \"octavia-b699-account-create-update-j7gtf\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.825035 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a25bce28-2142-4800-b3d6-e0059dae2e37-operator-scripts\") pod \"octavia-b699-account-create-update-j7gtf\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.826540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a25bce28-2142-4800-b3d6-e0059dae2e37-operator-scripts\") pod \"octavia-b699-account-create-update-j7gtf\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:53 crc kubenswrapper[4758]: I1203 18:37:53.846588 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4svl\" (UniqueName: \"kubernetes.io/projected/a25bce28-2142-4800-b3d6-e0059dae2e37-kube-api-access-p4svl\") pod \"octavia-b699-account-create-update-j7gtf\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.091653 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.253961 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5jrck" event={"ID":"b829c7e2-ce7c-46f9-830b-aa7227537d2c","Type":"ContainerStarted","Data":"02d99f9914021101c5f581027f208a553973dc1954ec88b5d2cd1fb1de42dcab"} Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.254002 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5jrck" event={"ID":"b829c7e2-ce7c-46f9-830b-aa7227537d2c","Type":"ContainerStarted","Data":"d090b9b6450a80bee8acab71f057761d3315f099c1a4c96cdb808b6a4ee7e3e0"} Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.255023 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.255053 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.256404 4758 generic.go:334] "Generic (PLEG): container finished" podID="412a365c-25b6-4058-9b1f-be4f76154b6a" containerID="c337c8fd45c5133eac7b793fb1afc59822251f795242a8372cacdebcba091999" exitCode=0 Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.257976 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lgqzq" event={"ID":"412a365c-25b6-4058-9b1f-be4f76154b6a","Type":"ContainerDied","Data":"c337c8fd45c5133eac7b793fb1afc59822251f795242a8372cacdebcba091999"} Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.258000 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lgqzq" event={"ID":"412a365c-25b6-4058-9b1f-be4f76154b6a","Type":"ContainerStarted","Data":"2fc21207a4fd2d3d41fb8de7b2a4170fa361ad82f3396b7f3bd12fcf20b9bba2"} Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.294914 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-5jrck" podStartSLOduration=4.294893005 podStartE2EDuration="4.294893005s" podCreationTimestamp="2025-12-03 18:37:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:37:54.285343367 +0000 UTC m=+6129.486720228" watchObservedRunningTime="2025-12-03 18:37:54.294893005 +0000 UTC m=+6129.496269866" Dec 03 18:37:54 crc kubenswrapper[4758]: I1203 18:37:54.578190 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-b699-account-create-update-j7gtf"] Dec 03 18:37:54 crc kubenswrapper[4758]: W1203 18:37:54.584371 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda25bce28_2142_4800_b3d6_e0059dae2e37.slice/crio-1d0905ba5c9d645ee5918e129b25fb3e953e406b4a5510484c09a63d764e313f WatchSource:0}: Error finding container 1d0905ba5c9d645ee5918e129b25fb3e953e406b4a5510484c09a63d764e313f: Status 404 returned error can't find the container with id 1d0905ba5c9d645ee5918e129b25fb3e953e406b4a5510484c09a63d764e313f Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.269604 4758 generic.go:334] "Generic (PLEG): container finished" podID="a25bce28-2142-4800-b3d6-e0059dae2e37" containerID="e4d51b7c62d37912e8953538b9be10602062551fee21f16298b3a19549882db7" exitCode=0 Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.269662 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-b699-account-create-update-j7gtf" event={"ID":"a25bce28-2142-4800-b3d6-e0059dae2e37","Type":"ContainerDied","Data":"e4d51b7c62d37912e8953538b9be10602062551fee21f16298b3a19549882db7"} Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.270134 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-b699-account-create-update-j7gtf" event={"ID":"a25bce28-2142-4800-b3d6-e0059dae2e37","Type":"ContainerStarted","Data":"1d0905ba5c9d645ee5918e129b25fb3e953e406b4a5510484c09a63d764e313f"} Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.651517 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.760663 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412a365c-25b6-4058-9b1f-be4f76154b6a-operator-scripts\") pod \"412a365c-25b6-4058-9b1f-be4f76154b6a\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.760782 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh7rs\" (UniqueName: \"kubernetes.io/projected/412a365c-25b6-4058-9b1f-be4f76154b6a-kube-api-access-mh7rs\") pod \"412a365c-25b6-4058-9b1f-be4f76154b6a\" (UID: \"412a365c-25b6-4058-9b1f-be4f76154b6a\") " Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.761510 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/412a365c-25b6-4058-9b1f-be4f76154b6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "412a365c-25b6-4058-9b1f-be4f76154b6a" (UID: "412a365c-25b6-4058-9b1f-be4f76154b6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.766400 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412a365c-25b6-4058-9b1f-be4f76154b6a-kube-api-access-mh7rs" (OuterVolumeSpecName: "kube-api-access-mh7rs") pod "412a365c-25b6-4058-9b1f-be4f76154b6a" (UID: "412a365c-25b6-4058-9b1f-be4f76154b6a"). InnerVolumeSpecName "kube-api-access-mh7rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.862449 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412a365c-25b6-4058-9b1f-be4f76154b6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:55 crc kubenswrapper[4758]: I1203 18:37:55.862487 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh7rs\" (UniqueName: \"kubernetes.io/projected/412a365c-25b6-4058-9b1f-be4f76154b6a-kube-api-access-mh7rs\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.278903 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lgqzq" event={"ID":"412a365c-25b6-4058-9b1f-be4f76154b6a","Type":"ContainerDied","Data":"2fc21207a4fd2d3d41fb8de7b2a4170fa361ad82f3396b7f3bd12fcf20b9bba2"} Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.279221 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fc21207a4fd2d3d41fb8de7b2a4170fa361ad82f3396b7f3bd12fcf20b9bba2" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.279006 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lgqzq" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.647840 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.681467 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4svl\" (UniqueName: \"kubernetes.io/projected/a25bce28-2142-4800-b3d6-e0059dae2e37-kube-api-access-p4svl\") pod \"a25bce28-2142-4800-b3d6-e0059dae2e37\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.681596 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a25bce28-2142-4800-b3d6-e0059dae2e37-operator-scripts\") pod \"a25bce28-2142-4800-b3d6-e0059dae2e37\" (UID: \"a25bce28-2142-4800-b3d6-e0059dae2e37\") " Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.682622 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a25bce28-2142-4800-b3d6-e0059dae2e37-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a25bce28-2142-4800-b3d6-e0059dae2e37" (UID: "a25bce28-2142-4800-b3d6-e0059dae2e37"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.689591 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a25bce28-2142-4800-b3d6-e0059dae2e37-kube-api-access-p4svl" (OuterVolumeSpecName: "kube-api-access-p4svl") pod "a25bce28-2142-4800-b3d6-e0059dae2e37" (UID: "a25bce28-2142-4800-b3d6-e0059dae2e37"). InnerVolumeSpecName "kube-api-access-p4svl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.783593 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4svl\" (UniqueName: \"kubernetes.io/projected/a25bce28-2142-4800-b3d6-e0059dae2e37-kube-api-access-p4svl\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:56 crc kubenswrapper[4758]: I1203 18:37:56.783636 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a25bce28-2142-4800-b3d6-e0059dae2e37-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:37:57 crc kubenswrapper[4758]: I1203 18:37:57.295222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-b699-account-create-update-j7gtf" event={"ID":"a25bce28-2142-4800-b3d6-e0059dae2e37","Type":"ContainerDied","Data":"1d0905ba5c9d645ee5918e129b25fb3e953e406b4a5510484c09a63d764e313f"} Dec 03 18:37:57 crc kubenswrapper[4758]: I1203 18:37:57.295271 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d0905ba5c9d645ee5918e129b25fb3e953e406b4a5510484c09a63d764e313f" Dec 03 18:37:57 crc kubenswrapper[4758]: I1203 18:37:57.295349 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b699-account-create-update-j7gtf" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.363011 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-w62m7"] Dec 03 18:37:59 crc kubenswrapper[4758]: E1203 18:37:59.363939 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a25bce28-2142-4800-b3d6-e0059dae2e37" containerName="mariadb-account-create-update" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.363952 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a25bce28-2142-4800-b3d6-e0059dae2e37" containerName="mariadb-account-create-update" Dec 03 18:37:59 crc kubenswrapper[4758]: E1203 18:37:59.363964 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412a365c-25b6-4058-9b1f-be4f76154b6a" containerName="mariadb-database-create" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.363972 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="412a365c-25b6-4058-9b1f-be4f76154b6a" containerName="mariadb-database-create" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.364160 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a25bce28-2142-4800-b3d6-e0059dae2e37" containerName="mariadb-account-create-update" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.364171 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="412a365c-25b6-4058-9b1f-be4f76154b6a" containerName="mariadb-database-create" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.365250 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.374167 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-w62m7"] Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.429209 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twfq6\" (UniqueName: \"kubernetes.io/projected/87c1fd37-adec-4fa9-b28c-9258306a3ad0-kube-api-access-twfq6\") pod \"octavia-persistence-db-create-w62m7\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.429394 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87c1fd37-adec-4fa9-b28c-9258306a3ad0-operator-scripts\") pod \"octavia-persistence-db-create-w62m7\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.530742 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87c1fd37-adec-4fa9-b28c-9258306a3ad0-operator-scripts\") pod \"octavia-persistence-db-create-w62m7\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.530841 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twfq6\" (UniqueName: \"kubernetes.io/projected/87c1fd37-adec-4fa9-b28c-9258306a3ad0-kube-api-access-twfq6\") pod \"octavia-persistence-db-create-w62m7\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.532015 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87c1fd37-adec-4fa9-b28c-9258306a3ad0-operator-scripts\") pod \"octavia-persistence-db-create-w62m7\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.561727 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twfq6\" (UniqueName: \"kubernetes.io/projected/87c1fd37-adec-4fa9-b28c-9258306a3ad0-kube-api-access-twfq6\") pod \"octavia-persistence-db-create-w62m7\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.695763 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.893618 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-a842-account-create-update-h2jjh"] Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.895191 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.897695 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.919713 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-a842-account-create-update-h2jjh"] Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.940508 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtxcz\" (UniqueName: \"kubernetes.io/projected/a288340d-898d-414d-beea-b40379693a89-kube-api-access-dtxcz\") pod \"octavia-a842-account-create-update-h2jjh\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:37:59 crc kubenswrapper[4758]: I1203 18:37:59.940977 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a288340d-898d-414d-beea-b40379693a89-operator-scripts\") pod \"octavia-a842-account-create-update-h2jjh\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.042765 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtxcz\" (UniqueName: \"kubernetes.io/projected/a288340d-898d-414d-beea-b40379693a89-kube-api-access-dtxcz\") pod \"octavia-a842-account-create-update-h2jjh\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.042898 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a288340d-898d-414d-beea-b40379693a89-operator-scripts\") pod \"octavia-a842-account-create-update-h2jjh\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.043961 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a288340d-898d-414d-beea-b40379693a89-operator-scripts\") pod \"octavia-a842-account-create-update-h2jjh\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.068990 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtxcz\" (UniqueName: \"kubernetes.io/projected/a288340d-898d-414d-beea-b40379693a89-kube-api-access-dtxcz\") pod \"octavia-a842-account-create-update-h2jjh\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.222061 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-w62m7"] Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.240006 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.337298 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-w62m7" event={"ID":"87c1fd37-adec-4fa9-b28c-9258306a3ad0","Type":"ContainerStarted","Data":"b3a5ddf136dae70cb2c078c7807cdf011fe5418c6bb6cf743475624ba1444a9b"} Dec 03 18:38:00 crc kubenswrapper[4758]: I1203 18:38:00.678621 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-a842-account-create-update-h2jjh"] Dec 03 18:38:00 crc kubenswrapper[4758]: W1203 18:38:00.691028 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda288340d_898d_414d_beea_b40379693a89.slice/crio-bb91917b1a9a4ef90be592c39d8c9f42dfb637ca83ba4141822dbd32f9047c60 WatchSource:0}: Error finding container bb91917b1a9a4ef90be592c39d8c9f42dfb637ca83ba4141822dbd32f9047c60: Status 404 returned error can't find the container with id bb91917b1a9a4ef90be592c39d8c9f42dfb637ca83ba4141822dbd32f9047c60 Dec 03 18:38:01 crc kubenswrapper[4758]: I1203 18:38:01.349243 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-a842-account-create-update-h2jjh" event={"ID":"a288340d-898d-414d-beea-b40379693a89","Type":"ContainerStarted","Data":"bb91917b1a9a4ef90be592c39d8c9f42dfb637ca83ba4141822dbd32f9047c60"} Dec 03 18:38:02 crc kubenswrapper[4758]: I1203 18:38:02.365500 4758 generic.go:334] "Generic (PLEG): container finished" podID="a288340d-898d-414d-beea-b40379693a89" containerID="7b5d0b89ed397f2b8c6f3b5e16fc897405d08d2a66d33867d44c7efb9e88722f" exitCode=0 Dec 03 18:38:02 crc kubenswrapper[4758]: I1203 18:38:02.365821 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-a842-account-create-update-h2jjh" event={"ID":"a288340d-898d-414d-beea-b40379693a89","Type":"ContainerDied","Data":"7b5d0b89ed397f2b8c6f3b5e16fc897405d08d2a66d33867d44c7efb9e88722f"} Dec 03 18:38:02 crc kubenswrapper[4758]: I1203 18:38:02.369139 4758 generic.go:334] "Generic (PLEG): container finished" podID="87c1fd37-adec-4fa9-b28c-9258306a3ad0" containerID="7bc767c44181b8918de1c95da92754abaacf5735a48379973f49890aec4ab3d2" exitCode=0 Dec 03 18:38:02 crc kubenswrapper[4758]: I1203 18:38:02.369189 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-w62m7" event={"ID":"87c1fd37-adec-4fa9-b28c-9258306a3ad0","Type":"ContainerDied","Data":"7bc767c44181b8918de1c95da92754abaacf5735a48379973f49890aec4ab3d2"} Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.864954 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.873529 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.913464 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87c1fd37-adec-4fa9-b28c-9258306a3ad0-operator-scripts\") pod \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.913531 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtxcz\" (UniqueName: \"kubernetes.io/projected/a288340d-898d-414d-beea-b40379693a89-kube-api-access-dtxcz\") pod \"a288340d-898d-414d-beea-b40379693a89\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.913575 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a288340d-898d-414d-beea-b40379693a89-operator-scripts\") pod \"a288340d-898d-414d-beea-b40379693a89\" (UID: \"a288340d-898d-414d-beea-b40379693a89\") " Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.913653 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twfq6\" (UniqueName: \"kubernetes.io/projected/87c1fd37-adec-4fa9-b28c-9258306a3ad0-kube-api-access-twfq6\") pod \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\" (UID: \"87c1fd37-adec-4fa9-b28c-9258306a3ad0\") " Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.914169 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87c1fd37-adec-4fa9-b28c-9258306a3ad0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "87c1fd37-adec-4fa9-b28c-9258306a3ad0" (UID: "87c1fd37-adec-4fa9-b28c-9258306a3ad0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.914305 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a288340d-898d-414d-beea-b40379693a89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a288340d-898d-414d-beea-b40379693a89" (UID: "a288340d-898d-414d-beea-b40379693a89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.921623 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a288340d-898d-414d-beea-b40379693a89-kube-api-access-dtxcz" (OuterVolumeSpecName: "kube-api-access-dtxcz") pod "a288340d-898d-414d-beea-b40379693a89" (UID: "a288340d-898d-414d-beea-b40379693a89"). InnerVolumeSpecName "kube-api-access-dtxcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:38:03 crc kubenswrapper[4758]: I1203 18:38:03.921875 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c1fd37-adec-4fa9-b28c-9258306a3ad0-kube-api-access-twfq6" (OuterVolumeSpecName: "kube-api-access-twfq6") pod "87c1fd37-adec-4fa9-b28c-9258306a3ad0" (UID: "87c1fd37-adec-4fa9-b28c-9258306a3ad0"). InnerVolumeSpecName "kube-api-access-twfq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.017079 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87c1fd37-adec-4fa9-b28c-9258306a3ad0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.017138 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtxcz\" (UniqueName: \"kubernetes.io/projected/a288340d-898d-414d-beea-b40379693a89-kube-api-access-dtxcz\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.017154 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a288340d-898d-414d-beea-b40379693a89-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.017191 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twfq6\" (UniqueName: \"kubernetes.io/projected/87c1fd37-adec-4fa9-b28c-9258306a3ad0-kube-api-access-twfq6\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.393812 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-w62m7" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.394080 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-w62m7" event={"ID":"87c1fd37-adec-4fa9-b28c-9258306a3ad0","Type":"ContainerDied","Data":"b3a5ddf136dae70cb2c078c7807cdf011fe5418c6bb6cf743475624ba1444a9b"} Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.394131 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3a5ddf136dae70cb2c078c7807cdf011fe5418c6bb6cf743475624ba1444a9b" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.396096 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-a842-account-create-update-h2jjh" event={"ID":"a288340d-898d-414d-beea-b40379693a89","Type":"ContainerDied","Data":"bb91917b1a9a4ef90be592c39d8c9f42dfb637ca83ba4141822dbd32f9047c60"} Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.396125 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb91917b1a9a4ef90be592c39d8c9f42dfb637ca83ba4141822dbd32f9047c60" Dec 03 18:38:04 crc kubenswrapper[4758]: I1203 18:38:04.396171 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-a842-account-create-update-h2jjh" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.483387 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-78cc97546b-97hmr"] Dec 03 18:38:10 crc kubenswrapper[4758]: E1203 18:38:10.484353 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c1fd37-adec-4fa9-b28c-9258306a3ad0" containerName="mariadb-database-create" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.484365 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c1fd37-adec-4fa9-b28c-9258306a3ad0" containerName="mariadb-database-create" Dec 03 18:38:10 crc kubenswrapper[4758]: E1203 18:38:10.484390 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a288340d-898d-414d-beea-b40379693a89" containerName="mariadb-account-create-update" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.484398 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a288340d-898d-414d-beea-b40379693a89" containerName="mariadb-account-create-update" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.484600 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c1fd37-adec-4fa9-b28c-9258306a3ad0" containerName="mariadb-database-create" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.484619 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a288340d-898d-414d-beea-b40379693a89" containerName="mariadb-account-create-update" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.485983 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.493595 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-78cc97546b-97hmr"] Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.496423 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-hxnvz" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.496842 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.497001 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.544366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-combined-ca-bundle\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.544474 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-octavia-run\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.545050 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-config-data-merged\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.545127 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-scripts\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.545160 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-config-data\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.646877 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-scripts\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.646932 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-config-data\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.647053 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-combined-ca-bundle\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.647153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-octavia-run\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.647606 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-octavia-run\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.647811 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-config-data-merged\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.648111 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-config-data-merged\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.654445 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-scripts\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.654644 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-config-data\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.657433 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1-combined-ca-bundle\") pod \"octavia-api-78cc97546b-97hmr\" (UID: \"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1\") " pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:10 crc kubenswrapper[4758]: I1203 18:38:10.817643 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:11 crc kubenswrapper[4758]: I1203 18:38:11.297908 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-78cc97546b-97hmr"] Dec 03 18:38:11 crc kubenswrapper[4758]: I1203 18:38:11.305987 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:38:11 crc kubenswrapper[4758]: I1203 18:38:11.465959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-78cc97546b-97hmr" event={"ID":"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1","Type":"ContainerStarted","Data":"5926ce53913baf3bc16ddf1241b1a2956c865ff90412ec7a6fd5116459195828"} Dec 03 18:38:13 crc kubenswrapper[4758]: I1203 18:38:13.196905 4758 scope.go:117] "RemoveContainer" containerID="dd15f4692cb8f037a9109df8374e6951f5eea9cdf439455644da5286f779d43c" Dec 03 18:38:13 crc kubenswrapper[4758]: I1203 18:38:13.242276 4758 scope.go:117] "RemoveContainer" containerID="353f25e9e0d987349abbc108e38be48376f64dac9ce271fe217f4bca1aee6390" Dec 03 18:38:13 crc kubenswrapper[4758]: I1203 18:38:13.294830 4758 scope.go:117] "RemoveContainer" containerID="b7da49a4b675859df73d5bd11925599345a99ffc4e7ac8dadfb365a87b1c418b" Dec 03 18:38:13 crc kubenswrapper[4758]: I1203 18:38:13.320806 4758 scope.go:117] "RemoveContainer" containerID="dc885c069b333d8660a1941fc0f26fe6797ee8d95297a047e5dac1ab6649b277" Dec 03 18:38:24 crc kubenswrapper[4758]: E1203 18:38:24.104288 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified" Dec 03 18:38:24 crc kubenswrapper[4758]: E1203 18:38:24.105409 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/container-scripts/init.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-merged,ReadOnly:false,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42437,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42437,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-api-78cc97546b-97hmr_openstack(9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 18:38:24 crc kubenswrapper[4758]: E1203 18:38:24.106872 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/octavia-api-78cc97546b-97hmr" podUID="9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1" Dec 03 18:38:24 crc kubenswrapper[4758]: E1203 18:38:24.600585 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified\\\"\"" pod="openstack/octavia-api-78cc97546b-97hmr" podUID="9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1" Dec 03 18:38:25 crc kubenswrapper[4758]: I1203 18:38:25.877974 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vts8k" podUID="45509ece-d0bc-4fed-aaea-825badb17447" containerName="ovn-controller" probeResult="failure" output=< Dec 03 18:38:25 crc kubenswrapper[4758]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 18:38:25 crc kubenswrapper[4758]: > Dec 03 18:38:25 crc kubenswrapper[4758]: I1203 18:38:25.884247 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:38:25 crc kubenswrapper[4758]: I1203 18:38:25.896980 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5jrck" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.035501 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vts8k-config-qmwds"] Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.036722 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.041789 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.055440 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vts8k-config-qmwds"] Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.059235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-scripts\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.059283 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkkpf\" (UniqueName: \"kubernetes.io/projected/d55d4422-e74c-458d-8a2a-543124a4da9e-kube-api-access-xkkpf\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.059401 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run-ovn\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.059431 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.059464 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-log-ovn\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.059498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-additional-scripts\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.160852 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run-ovn\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.160913 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.160960 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-log-ovn\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161004 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-additional-scripts\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161043 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-scripts\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161078 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkkpf\" (UniqueName: \"kubernetes.io/projected/d55d4422-e74c-458d-8a2a-543124a4da9e-kube-api-access-xkkpf\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161212 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run-ovn\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161233 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161457 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-log-ovn\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.161982 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-additional-scripts\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.163568 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-scripts\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.179018 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkkpf\" (UniqueName: \"kubernetes.io/projected/d55d4422-e74c-458d-8a2a-543124a4da9e-kube-api-access-xkkpf\") pod \"ovn-controller-vts8k-config-qmwds\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.369528 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:26 crc kubenswrapper[4758]: I1203 18:38:26.892269 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vts8k-config-qmwds"] Dec 03 18:38:27 crc kubenswrapper[4758]: I1203 18:38:27.627500 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vts8k-config-qmwds" event={"ID":"d55d4422-e74c-458d-8a2a-543124a4da9e","Type":"ContainerStarted","Data":"11008d53643cb88d1b75724952e6e3c8cb5ede96b320f8872a83dc341951c585"} Dec 03 18:38:27 crc kubenswrapper[4758]: I1203 18:38:27.627830 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vts8k-config-qmwds" event={"ID":"d55d4422-e74c-458d-8a2a-543124a4da9e","Type":"ContainerStarted","Data":"611efc30832f2dfb40df565aeb70d0ef0aa371acf675adba6db43420bda78a5f"} Dec 03 18:38:27 crc kubenswrapper[4758]: I1203 18:38:27.684485 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vts8k-config-qmwds" podStartSLOduration=1.684470278 podStartE2EDuration="1.684470278s" podCreationTimestamp="2025-12-03 18:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:38:27.682036802 +0000 UTC m=+6162.883413663" watchObservedRunningTime="2025-12-03 18:38:27.684470278 +0000 UTC m=+6162.885847139" Dec 03 18:38:28 crc kubenswrapper[4758]: E1203 18:38:28.367353 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd55d4422_e74c_458d_8a2a_543124a4da9e.slice/crio-conmon-11008d53643cb88d1b75724952e6e3c8cb5ede96b320f8872a83dc341951c585.scope\": RecentStats: unable to find data in memory cache]" Dec 03 18:38:28 crc kubenswrapper[4758]: I1203 18:38:28.637271 4758 generic.go:334] "Generic (PLEG): container finished" podID="d55d4422-e74c-458d-8a2a-543124a4da9e" containerID="11008d53643cb88d1b75724952e6e3c8cb5ede96b320f8872a83dc341951c585" exitCode=0 Dec 03 18:38:28 crc kubenswrapper[4758]: I1203 18:38:28.637311 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vts8k-config-qmwds" event={"ID":"d55d4422-e74c-458d-8a2a-543124a4da9e","Type":"ContainerDied","Data":"11008d53643cb88d1b75724952e6e3c8cb5ede96b320f8872a83dc341951c585"} Dec 03 18:38:29 crc kubenswrapper[4758]: I1203 18:38:29.993063 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.115014 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkkpf\" (UniqueName: \"kubernetes.io/projected/d55d4422-e74c-458d-8a2a-543124a4da9e-kube-api-access-xkkpf\") pod \"d55d4422-e74c-458d-8a2a-543124a4da9e\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.115059 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-scripts\") pod \"d55d4422-e74c-458d-8a2a-543124a4da9e\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.115090 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-additional-scripts\") pod \"d55d4422-e74c-458d-8a2a-543124a4da9e\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.115891 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d55d4422-e74c-458d-8a2a-543124a4da9e" (UID: "d55d4422-e74c-458d-8a2a-543124a4da9e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.116304 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-scripts" (OuterVolumeSpecName: "scripts") pod "d55d4422-e74c-458d-8a2a-543124a4da9e" (UID: "d55d4422-e74c-458d-8a2a-543124a4da9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.116446 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run\") pod \"d55d4422-e74c-458d-8a2a-543124a4da9e\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.116572 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-log-ovn\") pod \"d55d4422-e74c-458d-8a2a-543124a4da9e\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.116651 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run-ovn\") pod \"d55d4422-e74c-458d-8a2a-543124a4da9e\" (UID: \"d55d4422-e74c-458d-8a2a-543124a4da9e\") " Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.117016 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.117028 4758 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d55d4422-e74c-458d-8a2a-543124a4da9e-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.117052 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d55d4422-e74c-458d-8a2a-543124a4da9e" (UID: "d55d4422-e74c-458d-8a2a-543124a4da9e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.117068 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run" (OuterVolumeSpecName: "var-run") pod "d55d4422-e74c-458d-8a2a-543124a4da9e" (UID: "d55d4422-e74c-458d-8a2a-543124a4da9e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.117083 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d55d4422-e74c-458d-8a2a-543124a4da9e" (UID: "d55d4422-e74c-458d-8a2a-543124a4da9e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.120591 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d55d4422-e74c-458d-8a2a-543124a4da9e-kube-api-access-xkkpf" (OuterVolumeSpecName: "kube-api-access-xkkpf") pod "d55d4422-e74c-458d-8a2a-543124a4da9e" (UID: "d55d4422-e74c-458d-8a2a-543124a4da9e"). InnerVolumeSpecName "kube-api-access-xkkpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.219442 4758 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.219502 4758 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.219518 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkkpf\" (UniqueName: \"kubernetes.io/projected/d55d4422-e74c-458d-8a2a-543124a4da9e-kube-api-access-xkkpf\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.219532 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d55d4422-e74c-458d-8a2a-543124a4da9e-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.820104 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vts8k-config-qmwds" event={"ID":"d55d4422-e74c-458d-8a2a-543124a4da9e","Type":"ContainerDied","Data":"611efc30832f2dfb40df565aeb70d0ef0aa371acf675adba6db43420bda78a5f"} Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.820157 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="611efc30832f2dfb40df565aeb70d0ef0aa371acf675adba6db43420bda78a5f" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.820176 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vts8k-config-qmwds" Dec 03 18:38:30 crc kubenswrapper[4758]: I1203 18:38:30.872496 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vts8k" Dec 03 18:38:31 crc kubenswrapper[4758]: I1203 18:38:31.076493 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vts8k-config-qmwds"] Dec 03 18:38:31 crc kubenswrapper[4758]: I1203 18:38:31.087462 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vts8k-config-qmwds"] Dec 03 18:38:31 crc kubenswrapper[4758]: I1203 18:38:31.126471 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d55d4422-e74c-458d-8a2a-543124a4da9e" path="/var/lib/kubelet/pods/d55d4422-e74c-458d-8a2a-543124a4da9e/volumes" Dec 03 18:38:38 crc kubenswrapper[4758]: I1203 18:38:38.916342 4758 generic.go:334] "Generic (PLEG): container finished" podID="9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1" containerID="8ad37e8c45d9b6a51e9dbd52bfa079547179e2ee04565b83bfe6054682436574" exitCode=0 Dec 03 18:38:38 crc kubenswrapper[4758]: I1203 18:38:38.916417 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-78cc97546b-97hmr" event={"ID":"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1","Type":"ContainerDied","Data":"8ad37e8c45d9b6a51e9dbd52bfa079547179e2ee04565b83bfe6054682436574"} Dec 03 18:38:40 crc kubenswrapper[4758]: I1203 18:38:40.939040 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-78cc97546b-97hmr" event={"ID":"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1","Type":"ContainerStarted","Data":"657990e3bd5df4c94a5305552ca0286d0c54dce0c43486c4bd99ef94322e52e0"} Dec 03 18:38:40 crc kubenswrapper[4758]: I1203 18:38:40.939634 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-78cc97546b-97hmr" event={"ID":"9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1","Type":"ContainerStarted","Data":"2ec1ef54e49b2b19a6ff6b3efc9cac79d27d80b3cb8098156fd6eef5d095492f"} Dec 03 18:38:40 crc kubenswrapper[4758]: I1203 18:38:40.940950 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:40 crc kubenswrapper[4758]: I1203 18:38:40.940991 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:38:40 crc kubenswrapper[4758]: I1203 18:38:40.963385 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-78cc97546b-97hmr" podStartSLOduration=4.537266545 podStartE2EDuration="30.963366818s" podCreationTimestamp="2025-12-03 18:38:10 +0000 UTC" firstStartedPulling="2025-12-03 18:38:11.305725205 +0000 UTC m=+6146.507102066" lastFinishedPulling="2025-12-03 18:38:37.731825468 +0000 UTC m=+6172.933202339" observedRunningTime="2025-12-03 18:38:40.955410882 +0000 UTC m=+6176.156787763" watchObservedRunningTime="2025-12-03 18:38:40.963366818 +0000 UTC m=+6176.164743679" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.469089 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-9d5h7"] Dec 03 18:38:57 crc kubenswrapper[4758]: E1203 18:38:57.470088 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d55d4422-e74c-458d-8a2a-543124a4da9e" containerName="ovn-config" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.470105 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d55d4422-e74c-458d-8a2a-543124a4da9e" containerName="ovn-config" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.470334 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d55d4422-e74c-458d-8a2a-543124a4da9e" containerName="ovn-config" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.471620 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.483033 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.483175 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.483259 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.489401 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-9d5h7"] Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.565027 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0be9d87-ab3b-49f5-b103-7aa960127e4b-scripts\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.565112 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e0be9d87-ab3b-49f5-b103-7aa960127e4b-hm-ports\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.565414 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0be9d87-ab3b-49f5-b103-7aa960127e4b-config-data\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.565483 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e0be9d87-ab3b-49f5-b103-7aa960127e4b-config-data-merged\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.667733 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0be9d87-ab3b-49f5-b103-7aa960127e4b-config-data\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.667835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e0be9d87-ab3b-49f5-b103-7aa960127e4b-config-data-merged\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.668007 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0be9d87-ab3b-49f5-b103-7aa960127e4b-scripts\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.668149 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e0be9d87-ab3b-49f5-b103-7aa960127e4b-hm-ports\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.669284 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e0be9d87-ab3b-49f5-b103-7aa960127e4b-config-data-merged\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.670793 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e0be9d87-ab3b-49f5-b103-7aa960127e4b-hm-ports\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.674847 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0be9d87-ab3b-49f5-b103-7aa960127e4b-scripts\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.686142 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0be9d87-ab3b-49f5-b103-7aa960127e4b-config-data\") pod \"octavia-rsyslog-9d5h7\" (UID: \"e0be9d87-ab3b-49f5-b103-7aa960127e4b\") " pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:57 crc kubenswrapper[4758]: I1203 18:38:57.798100 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.075529 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fm79m"] Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.081404 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.084127 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.093313 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fm79m"] Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.180472 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5e7cbf33-7578-4233-b8ec-1d69624622f2-amphora-image\") pod \"octavia-image-upload-59f8cff499-fm79m\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.180576 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5e7cbf33-7578-4233-b8ec-1d69624622f2-httpd-config\") pod \"octavia-image-upload-59f8cff499-fm79m\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.282136 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5e7cbf33-7578-4233-b8ec-1d69624622f2-amphora-image\") pod \"octavia-image-upload-59f8cff499-fm79m\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.282227 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5e7cbf33-7578-4233-b8ec-1d69624622f2-httpd-config\") pod \"octavia-image-upload-59f8cff499-fm79m\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.283635 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5e7cbf33-7578-4233-b8ec-1d69624622f2-amphora-image\") pod \"octavia-image-upload-59f8cff499-fm79m\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.296244 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5e7cbf33-7578-4233-b8ec-1d69624622f2-httpd-config\") pod \"octavia-image-upload-59f8cff499-fm79m\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.375031 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-9d5h7"] Dec 03 18:38:58 crc kubenswrapper[4758]: W1203 18:38:58.381836 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0be9d87_ab3b_49f5_b103_7aa960127e4b.slice/crio-0dfaa4007cb2e9815b3a3db8a9df5406b614340c3c68c4f630c2cd925652e0d9 WatchSource:0}: Error finding container 0dfaa4007cb2e9815b3a3db8a9df5406b614340c3c68c4f630c2cd925652e0d9: Status 404 returned error can't find the container with id 0dfaa4007cb2e9815b3a3db8a9df5406b614340c3c68c4f630c2cd925652e0d9 Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.412545 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:38:58 crc kubenswrapper[4758]: I1203 18:38:58.875069 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fm79m"] Dec 03 18:38:58 crc kubenswrapper[4758]: W1203 18:38:58.877910 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e7cbf33_7578_4233_b8ec_1d69624622f2.slice/crio-00d2f7b53b7f00dd7e9c25ea5476a36a70081f92e42c4a4ae11af684b49f04c4 WatchSource:0}: Error finding container 00d2f7b53b7f00dd7e9c25ea5476a36a70081f92e42c4a4ae11af684b49f04c4: Status 404 returned error can't find the container with id 00d2f7b53b7f00dd7e9c25ea5476a36a70081f92e42c4a4ae11af684b49f04c4 Dec 03 18:38:59 crc kubenswrapper[4758]: I1203 18:38:59.131049 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9d5h7" event={"ID":"e0be9d87-ab3b-49f5-b103-7aa960127e4b","Type":"ContainerStarted","Data":"0dfaa4007cb2e9815b3a3db8a9df5406b614340c3c68c4f630c2cd925652e0d9"} Dec 03 18:38:59 crc kubenswrapper[4758]: I1203 18:38:59.131557 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fm79m" event={"ID":"5e7cbf33-7578-4233-b8ec-1d69624622f2","Type":"ContainerStarted","Data":"00d2f7b53b7f00dd7e9c25ea5476a36a70081f92e42c4a4ae11af684b49f04c4"} Dec 03 18:39:00 crc kubenswrapper[4758]: I1203 18:39:00.141074 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:39:00 crc kubenswrapper[4758]: I1203 18:39:00.327412 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-78cc97546b-97hmr" Dec 03 18:39:01 crc kubenswrapper[4758]: I1203 18:39:01.152945 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9d5h7" event={"ID":"e0be9d87-ab3b-49f5-b103-7aa960127e4b","Type":"ContainerStarted","Data":"754a0be2059cafa29ca060c5cfcbdc6d9872835ded5be88a49fff848a1cd3a6d"} Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.175524 4758 generic.go:334] "Generic (PLEG): container finished" podID="e0be9d87-ab3b-49f5-b103-7aa960127e4b" containerID="754a0be2059cafa29ca060c5cfcbdc6d9872835ded5be88a49fff848a1cd3a6d" exitCode=0 Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.175612 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9d5h7" event={"ID":"e0be9d87-ab3b-49f5-b103-7aa960127e4b","Type":"ContainerDied","Data":"754a0be2059cafa29ca060c5cfcbdc6d9872835ded5be88a49fff848a1cd3a6d"} Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.308443 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-gpjpc"] Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.310331 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.312499 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.322316 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-gpjpc"] Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.394321 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-combined-ca-bundle\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.394376 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data-merged\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.394465 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-scripts\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.394574 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.496787 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-scripts\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.496940 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.496990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-combined-ca-bundle\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.497023 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data-merged\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.497582 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data-merged\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.503041 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-scripts\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.510538 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-combined-ca-bundle\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.515263 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data\") pod \"octavia-db-sync-gpjpc\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:03 crc kubenswrapper[4758]: I1203 18:39:03.631297 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:04 crc kubenswrapper[4758]: W1203 18:39:04.108947 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddea3c817_521f_4709_b2c2_ab41fe7d9c81.slice/crio-a92b858069ae8cd79bc2907af48682f86425e616234f2407ce8931668bd50fb8 WatchSource:0}: Error finding container a92b858069ae8cd79bc2907af48682f86425e616234f2407ce8931668bd50fb8: Status 404 returned error can't find the container with id a92b858069ae8cd79bc2907af48682f86425e616234f2407ce8931668bd50fb8 Dec 03 18:39:04 crc kubenswrapper[4758]: I1203 18:39:04.113194 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-gpjpc"] Dec 03 18:39:04 crc kubenswrapper[4758]: I1203 18:39:04.185273 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gpjpc" event={"ID":"dea3c817-521f-4709-b2c2-ab41fe7d9c81","Type":"ContainerStarted","Data":"a92b858069ae8cd79bc2907af48682f86425e616234f2407ce8931668bd50fb8"} Dec 03 18:39:06 crc kubenswrapper[4758]: I1203 18:39:06.215392 4758 generic.go:334] "Generic (PLEG): container finished" podID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerID="d583bce5e20757c772d5d97f6b801869050141e60bbde59f7324fe621b40ecba" exitCode=0 Dec 03 18:39:06 crc kubenswrapper[4758]: I1203 18:39:06.215480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gpjpc" event={"ID":"dea3c817-521f-4709-b2c2-ab41fe7d9c81","Type":"ContainerDied","Data":"d583bce5e20757c772d5d97f6b801869050141e60bbde59f7324fe621b40ecba"} Dec 03 18:39:11 crc kubenswrapper[4758]: I1203 18:39:11.298574 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fm79m" event={"ID":"5e7cbf33-7578-4233-b8ec-1d69624622f2","Type":"ContainerStarted","Data":"bf8c51a8a57a2900634b2cd42da57c1dc77c90ebf2134f71bfd5a68da0cd7d66"} Dec 03 18:39:11 crc kubenswrapper[4758]: I1203 18:39:11.300833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gpjpc" event={"ID":"dea3c817-521f-4709-b2c2-ab41fe7d9c81","Type":"ContainerStarted","Data":"b281515c0cd165d6b3a31178a814fabae1924cebc0c8c28f3f840224fb5de799"} Dec 03 18:39:11 crc kubenswrapper[4758]: I1203 18:39:11.303035 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9d5h7" event={"ID":"e0be9d87-ab3b-49f5-b103-7aa960127e4b","Type":"ContainerStarted","Data":"a465e147f8c7c4d45140368bf674bf7189d587ef0b9b8c0f45e3a05419bcba51"} Dec 03 18:39:11 crc kubenswrapper[4758]: I1203 18:39:11.303322 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:39:11 crc kubenswrapper[4758]: I1203 18:39:11.349231 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-gpjpc" podStartSLOduration=8.34919772 podStartE2EDuration="8.34919772s" podCreationTimestamp="2025-12-03 18:39:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:39:11.339520099 +0000 UTC m=+6206.540897000" watchObservedRunningTime="2025-12-03 18:39:11.34919772 +0000 UTC m=+6206.550574631" Dec 03 18:39:11 crc kubenswrapper[4758]: I1203 18:39:11.369332 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-9d5h7" podStartSLOduration=1.837847877 podStartE2EDuration="14.369304615s" podCreationTimestamp="2025-12-03 18:38:57 +0000 UTC" firstStartedPulling="2025-12-03 18:38:58.386359989 +0000 UTC m=+6193.587736850" lastFinishedPulling="2025-12-03 18:39:10.917816737 +0000 UTC m=+6206.119193588" observedRunningTime="2025-12-03 18:39:11.358266366 +0000 UTC m=+6206.559643227" watchObservedRunningTime="2025-12-03 18:39:11.369304615 +0000 UTC m=+6206.570681506" Dec 03 18:39:12 crc kubenswrapper[4758]: I1203 18:39:12.314134 4758 generic.go:334] "Generic (PLEG): container finished" podID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerID="bf8c51a8a57a2900634b2cd42da57c1dc77c90ebf2134f71bfd5a68da0cd7d66" exitCode=0 Dec 03 18:39:12 crc kubenswrapper[4758]: I1203 18:39:12.314188 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fm79m" event={"ID":"5e7cbf33-7578-4233-b8ec-1d69624622f2","Type":"ContainerDied","Data":"bf8c51a8a57a2900634b2cd42da57c1dc77c90ebf2134f71bfd5a68da0cd7d66"} Dec 03 18:39:18 crc kubenswrapper[4758]: I1203 18:39:18.382722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fm79m" event={"ID":"5e7cbf33-7578-4233-b8ec-1d69624622f2","Type":"ContainerStarted","Data":"b15f82c5a42cc186b5533a50df8e62c3aa96cb092f6861eb5f50b22afc40b101"} Dec 03 18:39:18 crc kubenswrapper[4758]: I1203 18:39:18.414895 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-fm79m" podStartSLOduration=1.8884708510000001 podStartE2EDuration="20.414867789s" podCreationTimestamp="2025-12-03 18:38:58 +0000 UTC" firstStartedPulling="2025-12-03 18:38:58.880267846 +0000 UTC m=+6194.081644707" lastFinishedPulling="2025-12-03 18:39:17.406664794 +0000 UTC m=+6212.608041645" observedRunningTime="2025-12-03 18:39:18.40198693 +0000 UTC m=+6213.603363851" watchObservedRunningTime="2025-12-03 18:39:18.414867789 +0000 UTC m=+6213.616244690" Dec 03 18:39:19 crc kubenswrapper[4758]: I1203 18:39:19.395785 4758 generic.go:334] "Generic (PLEG): container finished" podID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerID="b281515c0cd165d6b3a31178a814fabae1924cebc0c8c28f3f840224fb5de799" exitCode=0 Dec 03 18:39:19 crc kubenswrapper[4758]: I1203 18:39:19.395877 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gpjpc" event={"ID":"dea3c817-521f-4709-b2c2-ab41fe7d9c81","Type":"ContainerDied","Data":"b281515c0cd165d6b3a31178a814fabae1924cebc0c8c28f3f840224fb5de799"} Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.846425 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.968649 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data-merged\") pod \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.968807 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-combined-ca-bundle\") pod \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.968975 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data\") pod \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.969251 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-scripts\") pod \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\" (UID: \"dea3c817-521f-4709-b2c2-ab41fe7d9c81\") " Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.974857 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-scripts" (OuterVolumeSpecName: "scripts") pod "dea3c817-521f-4709-b2c2-ab41fe7d9c81" (UID: "dea3c817-521f-4709-b2c2-ab41fe7d9c81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.975205 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data" (OuterVolumeSpecName: "config-data") pod "dea3c817-521f-4709-b2c2-ab41fe7d9c81" (UID: "dea3c817-521f-4709-b2c2-ab41fe7d9c81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:39:20 crc kubenswrapper[4758]: I1203 18:39:20.993142 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "dea3c817-521f-4709-b2c2-ab41fe7d9c81" (UID: "dea3c817-521f-4709-b2c2-ab41fe7d9c81"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.004589 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dea3c817-521f-4709-b2c2-ab41fe7d9c81" (UID: "dea3c817-521f-4709-b2c2-ab41fe7d9c81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.071615 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data-merged\") on node \"crc\" DevicePath \"\"" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.071894 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.071967 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.072058 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea3c817-521f-4709-b2c2-ab41fe7d9c81-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.422782 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gpjpc" event={"ID":"dea3c817-521f-4709-b2c2-ab41fe7d9c81","Type":"ContainerDied","Data":"a92b858069ae8cd79bc2907af48682f86425e616234f2407ce8931668bd50fb8"} Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.422825 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a92b858069ae8cd79bc2907af48682f86425e616234f2407ce8931668bd50fb8" Dec 03 18:39:21 crc kubenswrapper[4758]: I1203 18:39:21.422859 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gpjpc" Dec 03 18:39:27 crc kubenswrapper[4758]: I1203 18:39:27.827830 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-9d5h7" Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.268921 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fm79m"] Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.269941 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-fm79m" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerName="octavia-amphora-httpd" containerID="cri-o://b15f82c5a42cc186b5533a50df8e62c3aa96cb092f6861eb5f50b22afc40b101" gracePeriod=30 Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.692103 4758 generic.go:334] "Generic (PLEG): container finished" podID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerID="b15f82c5a42cc186b5533a50df8e62c3aa96cb092f6861eb5f50b22afc40b101" exitCode=0 Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.692172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fm79m" event={"ID":"5e7cbf33-7578-4233-b8ec-1d69624622f2","Type":"ContainerDied","Data":"b15f82c5a42cc186b5533a50df8e62c3aa96cb092f6861eb5f50b22afc40b101"} Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.867966 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.935870 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5e7cbf33-7578-4233-b8ec-1d69624622f2-httpd-config\") pod \"5e7cbf33-7578-4233-b8ec-1d69624622f2\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.936232 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5e7cbf33-7578-4233-b8ec-1d69624622f2-amphora-image\") pod \"5e7cbf33-7578-4233-b8ec-1d69624622f2\" (UID: \"5e7cbf33-7578-4233-b8ec-1d69624622f2\") " Dec 03 18:39:43 crc kubenswrapper[4758]: I1203 18:39:43.966616 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7cbf33-7578-4233-b8ec-1d69624622f2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5e7cbf33-7578-4233-b8ec-1d69624622f2" (UID: "5e7cbf33-7578-4233-b8ec-1d69624622f2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.017819 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7cbf33-7578-4233-b8ec-1d69624622f2-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "5e7cbf33-7578-4233-b8ec-1d69624622f2" (UID: "5e7cbf33-7578-4233-b8ec-1d69624622f2"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.040084 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5e7cbf33-7578-4233-b8ec-1d69624622f2-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.040129 4758 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5e7cbf33-7578-4233-b8ec-1d69624622f2-amphora-image\") on node \"crc\" DevicePath \"\"" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.704765 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-fm79m" event={"ID":"5e7cbf33-7578-4233-b8ec-1d69624622f2","Type":"ContainerDied","Data":"00d2f7b53b7f00dd7e9c25ea5476a36a70081f92e42c4a4ae11af684b49f04c4"} Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.705072 4758 scope.go:117] "RemoveContainer" containerID="b15f82c5a42cc186b5533a50df8e62c3aa96cb092f6861eb5f50b22afc40b101" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.704822 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-fm79m" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.735905 4758 scope.go:117] "RemoveContainer" containerID="bf8c51a8a57a2900634b2cd42da57c1dc77c90ebf2134f71bfd5a68da0cd7d66" Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.741811 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fm79m"] Dec 03 18:39:44 crc kubenswrapper[4758]: I1203 18:39:44.752959 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-fm79m"] Dec 03 18:39:45 crc kubenswrapper[4758]: I1203 18:39:45.125329 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" path="/var/lib/kubelet/pods/5e7cbf33-7578-4233-b8ec-1d69624622f2/volumes" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.544139 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-mfngn"] Dec 03 18:39:47 crc kubenswrapper[4758]: E1203 18:39:47.545203 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerName="init" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.545219 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerName="init" Dec 03 18:39:47 crc kubenswrapper[4758]: E1203 18:39:47.545252 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerName="init" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.545260 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerName="init" Dec 03 18:39:47 crc kubenswrapper[4758]: E1203 18:39:47.545275 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerName="octavia-amphora-httpd" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.545285 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerName="octavia-amphora-httpd" Dec 03 18:39:47 crc kubenswrapper[4758]: E1203 18:39:47.545303 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerName="octavia-db-sync" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.545310 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerName="octavia-db-sync" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.545545 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" containerName="octavia-db-sync" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.545576 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7cbf33-7578-4233-b8ec-1d69624622f2" containerName="octavia-amphora-httpd" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.546847 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.552337 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.556102 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-mfngn"] Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.622385 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/397e1e84-cff5-44c1-9ce9-1ba6022b95db-httpd-config\") pod \"octavia-image-upload-59f8cff499-mfngn\" (UID: \"397e1e84-cff5-44c1-9ce9-1ba6022b95db\") " pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.622576 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/397e1e84-cff5-44c1-9ce9-1ba6022b95db-amphora-image\") pod \"octavia-image-upload-59f8cff499-mfngn\" (UID: \"397e1e84-cff5-44c1-9ce9-1ba6022b95db\") " pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.723505 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/397e1e84-cff5-44c1-9ce9-1ba6022b95db-amphora-image\") pod \"octavia-image-upload-59f8cff499-mfngn\" (UID: \"397e1e84-cff5-44c1-9ce9-1ba6022b95db\") " pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.723576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/397e1e84-cff5-44c1-9ce9-1ba6022b95db-httpd-config\") pod \"octavia-image-upload-59f8cff499-mfngn\" (UID: \"397e1e84-cff5-44c1-9ce9-1ba6022b95db\") " pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.724280 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/397e1e84-cff5-44c1-9ce9-1ba6022b95db-amphora-image\") pod \"octavia-image-upload-59f8cff499-mfngn\" (UID: \"397e1e84-cff5-44c1-9ce9-1ba6022b95db\") " pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.737716 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/397e1e84-cff5-44c1-9ce9-1ba6022b95db-httpd-config\") pod \"octavia-image-upload-59f8cff499-mfngn\" (UID: \"397e1e84-cff5-44c1-9ce9-1ba6022b95db\") " pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:47 crc kubenswrapper[4758]: I1203 18:39:47.870485 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-mfngn" Dec 03 18:39:48 crc kubenswrapper[4758]: I1203 18:39:48.341434 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-mfngn"] Dec 03 18:39:48 crc kubenswrapper[4758]: I1203 18:39:48.753884 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-mfngn" event={"ID":"397e1e84-cff5-44c1-9ce9-1ba6022b95db","Type":"ContainerStarted","Data":"f071466f1318795e62e34f88df5b75fd096bf2fb83b82f2ad199ca1d2f1572dc"} Dec 03 18:39:49 crc kubenswrapper[4758]: I1203 18:39:49.767242 4758 generic.go:334] "Generic (PLEG): container finished" podID="397e1e84-cff5-44c1-9ce9-1ba6022b95db" containerID="7e6d3062fd42cfc71fc50e8cb0c52f253eaa8970297cafa3b06c6991a98dd20b" exitCode=0 Dec 03 18:39:49 crc kubenswrapper[4758]: I1203 18:39:49.767325 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-mfngn" event={"ID":"397e1e84-cff5-44c1-9ce9-1ba6022b95db","Type":"ContainerDied","Data":"7e6d3062fd42cfc71fc50e8cb0c52f253eaa8970297cafa3b06c6991a98dd20b"} Dec 03 18:39:51 crc kubenswrapper[4758]: I1203 18:39:51.785899 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-mfngn" event={"ID":"397e1e84-cff5-44c1-9ce9-1ba6022b95db","Type":"ContainerStarted","Data":"751e4bc746fe1949303a811979a02404fadc9f556d1e909fdea8d6208e6d94c4"} Dec 03 18:39:51 crc kubenswrapper[4758]: I1203 18:39:51.812693 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-mfngn" podStartSLOduration=2.179795297 podStartE2EDuration="4.812652282s" podCreationTimestamp="2025-12-03 18:39:47 +0000 UTC" firstStartedPulling="2025-12-03 18:39:48.344790583 +0000 UTC m=+6243.546167454" lastFinishedPulling="2025-12-03 18:39:50.977647578 +0000 UTC m=+6246.179024439" observedRunningTime="2025-12-03 18:39:51.803518064 +0000 UTC m=+6247.004894925" watchObservedRunningTime="2025-12-03 18:39:51.812652282 +0000 UTC m=+6247.014029153" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.842187 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-8wngx"] Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.845782 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.852173 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.852390 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.852743 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.855457 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-8wngx"] Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.960538 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8d1e1443-d203-4e87-8a80-f2d9785eb981-config-data-merged\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.961014 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-scripts\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.961078 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-combined-ca-bundle\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.961336 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-config-data\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.961481 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-amphora-certs\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:04 crc kubenswrapper[4758]: I1203 18:40:04.961603 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8d1e1443-d203-4e87-8a80-f2d9785eb981-hm-ports\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.048946 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5a20-account-create-update-wd9nn"] Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.062113 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-kf7sz"] Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.063205 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-config-data\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.063265 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-amphora-certs\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.063294 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8d1e1443-d203-4e87-8a80-f2d9785eb981-hm-ports\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.063346 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8d1e1443-d203-4e87-8a80-f2d9785eb981-config-data-merged\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.063365 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-scripts\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.063393 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-combined-ca-bundle\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.064815 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8d1e1443-d203-4e87-8a80-f2d9785eb981-config-data-merged\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.065444 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8d1e1443-d203-4e87-8a80-f2d9785eb981-hm-ports\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.069478 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-amphora-certs\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.070314 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-scripts\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.070458 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-combined-ca-bundle\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.070832 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e1443-d203-4e87-8a80-f2d9785eb981-config-data\") pod \"octavia-healthmanager-8wngx\" (UID: \"8d1e1443-d203-4e87-8a80-f2d9785eb981\") " pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.071416 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-kf7sz"] Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.079196 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5a20-account-create-update-wd9nn"] Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.126705 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ddf7b6b-9452-49ba-b875-6caf9392059f" path="/var/lib/kubelet/pods/4ddf7b6b-9452-49ba-b875-6caf9392059f/volumes" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.127515 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="804c639b-8c29-4d6c-864a-77715bf8b004" path="/var/lib/kubelet/pods/804c639b-8c29-4d6c-864a-77715bf8b004/volumes" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.172404 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.718197 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-8wngx"] Dec 03 18:40:05 crc kubenswrapper[4758]: W1203 18:40:05.728607 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d1e1443_d203_4e87_8a80_f2d9785eb981.slice/crio-e9fdf5047d86b26671093e61e61255ef7a1fddff3d4d0d7a8da1dafb4071c13f WatchSource:0}: Error finding container e9fdf5047d86b26671093e61e61255ef7a1fddff3d4d0d7a8da1dafb4071c13f: Status 404 returned error can't find the container with id e9fdf5047d86b26671093e61e61255ef7a1fddff3d4d0d7a8da1dafb4071c13f Dec 03 18:40:05 crc kubenswrapper[4758]: I1203 18:40:05.942277 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-8wngx" event={"ID":"8d1e1443-d203-4e87-8a80-f2d9785eb981","Type":"ContainerStarted","Data":"e9fdf5047d86b26671093e61e61255ef7a1fddff3d4d0d7a8da1dafb4071c13f"} Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.361363 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-tdpvl"] Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.363180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.365592 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.365598 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.381580 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-tdpvl"] Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.490086 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-amphora-certs\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.490148 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-combined-ca-bundle\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.490315 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-hm-ports\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.490447 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-config-data-merged\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.490589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-scripts\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.490745 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-config-data\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.593071 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-amphora-certs\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.593123 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-combined-ca-bundle\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.593156 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-hm-ports\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.593191 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-config-data-merged\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.593226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-scripts\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.593526 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-config-data\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.594266 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-config-data-merged\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.595255 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-hm-ports\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.599995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-config-data\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.599995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-combined-ca-bundle\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.600425 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-scripts\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.600888 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907-amphora-certs\") pod \"octavia-housekeeping-tdpvl\" (UID: \"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907\") " pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.679347 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:06 crc kubenswrapper[4758]: I1203 18:40:06.968450 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-8wngx" event={"ID":"8d1e1443-d203-4e87-8a80-f2d9785eb981","Type":"ContainerStarted","Data":"da1366d28bc13650362f44dd65c0c962463fdab69d0c1115a8216f9f2d09b5a1"} Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.061360 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-tdpvl"] Dec 03 18:40:07 crc kubenswrapper[4758]: W1203 18:40:07.066792 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2a8136_2ce0_4cb9_9bae_b2ea4dbd3907.slice/crio-e1a260c7214327338c9cdc22fdc9fdeb86322570d313805c304d619f88a9f03b WatchSource:0}: Error finding container e1a260c7214327338c9cdc22fdc9fdeb86322570d313805c304d619f88a9f03b: Status 404 returned error can't find the container with id e1a260c7214327338c9cdc22fdc9fdeb86322570d313805c304d619f88a9f03b Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.530472 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-l9dgm"] Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.536886 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.539609 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.539698 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.544927 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-l9dgm"] Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.612355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e4c53e2b-203e-4767-9943-d063ef13ff7c-config-data-merged\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.612616 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-combined-ca-bundle\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.612660 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e4c53e2b-203e-4767-9943-d063ef13ff7c-hm-ports\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.612890 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-amphora-certs\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.612990 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-config-data\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.613078 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-scripts\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e4c53e2b-203e-4767-9943-d063ef13ff7c-config-data-merged\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714520 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-combined-ca-bundle\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714573 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e4c53e2b-203e-4767-9943-d063ef13ff7c-hm-ports\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714635 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-amphora-certs\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714696 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-config-data\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714743 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-scripts\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.714988 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e4c53e2b-203e-4767-9943-d063ef13ff7c-config-data-merged\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.716011 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e4c53e2b-203e-4767-9943-d063ef13ff7c-hm-ports\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.721029 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-amphora-certs\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.721510 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-scripts\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.722634 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-combined-ca-bundle\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.744819 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c53e2b-203e-4767-9943-d063ef13ff7c-config-data\") pod \"octavia-worker-l9dgm\" (UID: \"e4c53e2b-203e-4767-9943-d063ef13ff7c\") " pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.870870 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.987574 4758 generic.go:334] "Generic (PLEG): container finished" podID="8d1e1443-d203-4e87-8a80-f2d9785eb981" containerID="da1366d28bc13650362f44dd65c0c962463fdab69d0c1115a8216f9f2d09b5a1" exitCode=0 Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.987648 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-8wngx" event={"ID":"8d1e1443-d203-4e87-8a80-f2d9785eb981","Type":"ContainerDied","Data":"da1366d28bc13650362f44dd65c0c962463fdab69d0c1115a8216f9f2d09b5a1"} Dec 03 18:40:07 crc kubenswrapper[4758]: I1203 18:40:07.989520 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-tdpvl" event={"ID":"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907","Type":"ContainerStarted","Data":"e1a260c7214327338c9cdc22fdc9fdeb86322570d313805c304d619f88a9f03b"} Dec 03 18:40:08 crc kubenswrapper[4758]: I1203 18:40:08.467714 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-l9dgm"] Dec 03 18:40:08 crc kubenswrapper[4758]: W1203 18:40:08.729974 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4c53e2b_203e_4767_9943_d063ef13ff7c.slice/crio-b69c58bdc202fd65d4e160fdb8c28bbec3dc7776723588cb3b027360d46f2a9b WatchSource:0}: Error finding container b69c58bdc202fd65d4e160fdb8c28bbec3dc7776723588cb3b027360d46f2a9b: Status 404 returned error can't find the container with id b69c58bdc202fd65d4e160fdb8c28bbec3dc7776723588cb3b027360d46f2a9b Dec 03 18:40:09 crc kubenswrapper[4758]: I1203 18:40:09.003024 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-8wngx" event={"ID":"8d1e1443-d203-4e87-8a80-f2d9785eb981","Type":"ContainerStarted","Data":"c2424599bd48baa27278137ce408fde047d00036b1ea77b2f92a53f67974e25e"} Dec 03 18:40:09 crc kubenswrapper[4758]: I1203 18:40:09.004311 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:09 crc kubenswrapper[4758]: I1203 18:40:09.005637 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-l9dgm" event={"ID":"e4c53e2b-203e-4767-9943-d063ef13ff7c","Type":"ContainerStarted","Data":"b69c58bdc202fd65d4e160fdb8c28bbec3dc7776723588cb3b027360d46f2a9b"} Dec 03 18:40:09 crc kubenswrapper[4758]: I1203 18:40:09.031834 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-8wngx" podStartSLOduration=5.031816734 podStartE2EDuration="5.031816734s" podCreationTimestamp="2025-12-03 18:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:40:09.021451663 +0000 UTC m=+6264.222828534" watchObservedRunningTime="2025-12-03 18:40:09.031816734 +0000 UTC m=+6264.233193595" Dec 03 18:40:10 crc kubenswrapper[4758]: I1203 18:40:10.017668 4758 generic.go:334] "Generic (PLEG): container finished" podID="1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907" containerID="6e07a36a273c7a9a4e3988157c8e607f987b7fa601701d708ee61ba3c264b677" exitCode=0 Dec 03 18:40:10 crc kubenswrapper[4758]: I1203 18:40:10.018196 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-tdpvl" event={"ID":"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907","Type":"ContainerDied","Data":"6e07a36a273c7a9a4e3988157c8e607f987b7fa601701d708ee61ba3c264b677"} Dec 03 18:40:11 crc kubenswrapper[4758]: I1203 18:40:11.034421 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-tdpvl" event={"ID":"1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907","Type":"ContainerStarted","Data":"cb34e09d0f7eda0cab9ed58ae72bf250f3391b5bcf1d6c914ab830bfc1ba9ca3"} Dec 03 18:40:11 crc kubenswrapper[4758]: I1203 18:40:11.035053 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:11 crc kubenswrapper[4758]: I1203 18:40:11.063228 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-tdpvl" podStartSLOduration=3.357599417 podStartE2EDuration="5.0632138s" podCreationTimestamp="2025-12-03 18:40:06 +0000 UTC" firstStartedPulling="2025-12-03 18:40:07.069040296 +0000 UTC m=+6262.270417157" lastFinishedPulling="2025-12-03 18:40:08.774654679 +0000 UTC m=+6263.976031540" observedRunningTime="2025-12-03 18:40:11.062229163 +0000 UTC m=+6266.263606014" watchObservedRunningTime="2025-12-03 18:40:11.0632138 +0000 UTC m=+6266.264590661" Dec 03 18:40:11 crc kubenswrapper[4758]: I1203 18:40:11.394356 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:40:11 crc kubenswrapper[4758]: I1203 18:40:11.394407 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:40:12 crc kubenswrapper[4758]: I1203 18:40:12.041832 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-v9vjh"] Dec 03 18:40:12 crc kubenswrapper[4758]: I1203 18:40:12.044795 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-l9dgm" event={"ID":"e4c53e2b-203e-4767-9943-d063ef13ff7c","Type":"ContainerStarted","Data":"8fc43d60e9e646ae29552afc7a9de33f52e6e2023f3cbb0c57850b44a252917f"} Dec 03 18:40:12 crc kubenswrapper[4758]: I1203 18:40:12.055212 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-v9vjh"] Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.053809 4758 generic.go:334] "Generic (PLEG): container finished" podID="e4c53e2b-203e-4767-9943-d063ef13ff7c" containerID="8fc43d60e9e646ae29552afc7a9de33f52e6e2023f3cbb0c57850b44a252917f" exitCode=0 Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.053898 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-l9dgm" event={"ID":"e4c53e2b-203e-4767-9943-d063ef13ff7c","Type":"ContainerDied","Data":"8fc43d60e9e646ae29552afc7a9de33f52e6e2023f3cbb0c57850b44a252917f"} Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.127164 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9547728c-ddd1-4caa-b665-b584c15304ac" path="/var/lib/kubelet/pods/9547728c-ddd1-4caa-b665-b584c15304ac/volumes" Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.473601 4758 scope.go:117] "RemoveContainer" containerID="10f5969c64aa23c105137b7c839a8316a6ac45aa1709a9d7cf6015009416420a" Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.510264 4758 scope.go:117] "RemoveContainer" containerID="1f9beb59e1ebac57a7b7f152c1626cc803ccccf03c7475d421471baaa2aa1c48" Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.555479 4758 scope.go:117] "RemoveContainer" containerID="90dfacb6f481088b94e0cee74f4e22f3ecbc31330fd7af7f2df0c004833529f3" Dec 03 18:40:13 crc kubenswrapper[4758]: I1203 18:40:13.613778 4758 scope.go:117] "RemoveContainer" containerID="ccb91b0ccb8d85f672dcf295e826ea2f46440636f4cdd12ce24e4e938b942dcf" Dec 03 18:40:14 crc kubenswrapper[4758]: I1203 18:40:14.080363 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-l9dgm" event={"ID":"e4c53e2b-203e-4767-9943-d063ef13ff7c","Type":"ContainerStarted","Data":"b42ebd734bb9b46e7af0743e5e3ba96c1a377a26cf8777e5cf86a7f4196259ee"} Dec 03 18:40:14 crc kubenswrapper[4758]: I1203 18:40:14.080533 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:14 crc kubenswrapper[4758]: I1203 18:40:14.120022 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-l9dgm" podStartSLOduration=4.76145545 podStartE2EDuration="7.119998066s" podCreationTimestamp="2025-12-03 18:40:07 +0000 UTC" firstStartedPulling="2025-12-03 18:40:08.732632491 +0000 UTC m=+6263.934009342" lastFinishedPulling="2025-12-03 18:40:11.091175097 +0000 UTC m=+6266.292551958" observedRunningTime="2025-12-03 18:40:14.111768183 +0000 UTC m=+6269.313145054" watchObservedRunningTime="2025-12-03 18:40:14.119998066 +0000 UTC m=+6269.321374927" Dec 03 18:40:20 crc kubenswrapper[4758]: I1203 18:40:20.201441 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-8wngx" Dec 03 18:40:21 crc kubenswrapper[4758]: I1203 18:40:21.715943 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-tdpvl" Dec 03 18:40:22 crc kubenswrapper[4758]: I1203 18:40:22.905605 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-l9dgm" Dec 03 18:40:38 crc kubenswrapper[4758]: I1203 18:40:38.054580 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-da08-account-create-update-ts4jj"] Dec 03 18:40:38 crc kubenswrapper[4758]: I1203 18:40:38.068291 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pfhjl"] Dec 03 18:40:38 crc kubenswrapper[4758]: I1203 18:40:38.081424 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-da08-account-create-update-ts4jj"] Dec 03 18:40:38 crc kubenswrapper[4758]: I1203 18:40:38.093829 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pfhjl"] Dec 03 18:40:39 crc kubenswrapper[4758]: I1203 18:40:39.133113 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35fa890-62ae-4f36-84d4-44cb9dafde45" path="/var/lib/kubelet/pods/c35fa890-62ae-4f36-84d4-44cb9dafde45/volumes" Dec 03 18:40:39 crc kubenswrapper[4758]: I1203 18:40:39.134436 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1dd1a04-1178-4995-9efa-d843f24e8700" path="/var/lib/kubelet/pods/d1dd1a04-1178-4995-9efa-d843f24e8700/volumes" Dec 03 18:40:41 crc kubenswrapper[4758]: I1203 18:40:41.394455 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:40:41 crc kubenswrapper[4758]: I1203 18:40:41.394518 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:40:47 crc kubenswrapper[4758]: I1203 18:40:47.045929 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9gb7v"] Dec 03 18:40:47 crc kubenswrapper[4758]: I1203 18:40:47.055762 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9gb7v"] Dec 03 18:40:47 crc kubenswrapper[4758]: I1203 18:40:47.124879 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76331360-c8e3-48e2-a71d-1313e699ecd9" path="/var/lib/kubelet/pods/76331360-c8e3-48e2-a71d-1313e699ecd9/volumes" Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.394418 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.395035 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.395077 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.395847 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"55064fb8d522abc0cbfb578b1ed4ff679982f3cf1154c3a5e9439b6fbc880f07"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.395900 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://55064fb8d522abc0cbfb578b1ed4ff679982f3cf1154c3a5e9439b6fbc880f07" gracePeriod=600 Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.620514 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="55064fb8d522abc0cbfb578b1ed4ff679982f3cf1154c3a5e9439b6fbc880f07" exitCode=0 Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.620561 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"55064fb8d522abc0cbfb578b1ed4ff679982f3cf1154c3a5e9439b6fbc880f07"} Dec 03 18:41:11 crc kubenswrapper[4758]: I1203 18:41:11.620596 4758 scope.go:117] "RemoveContainer" containerID="4b4b7a4c76ec7bb30a96b9a0453b919e87e43bc1961dfd9be5f1fbf06ac515d3" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.636717 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00"} Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.707763 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7746fd787c-t4ld6"] Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.709797 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.715377 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-hqjmx" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.715533 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.715775 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.715472 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.722943 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7746fd787c-t4ld6"] Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.755303 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.755625 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-log" containerID="cri-o://303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38" gracePeriod=30 Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.755730 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-httpd" containerID="cri-o://0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179" gracePeriod=30 Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.817814 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74c78b48d9-82vjv"] Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.819332 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.836106 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74c78b48d9-82vjv"] Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.877922 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12d17af-6af8-402d-86a3-65079d25114b-logs\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.877994 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-scripts\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.878052 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p4m4\" (UniqueName: \"kubernetes.io/projected/f12d17af-6af8-402d-86a3-65079d25114b-kube-api-access-2p4m4\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.878151 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-config-data\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.878180 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f12d17af-6af8-402d-86a3-65079d25114b-horizon-secret-key\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.887372 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.887674 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-log" containerID="cri-o://17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98" gracePeriod=30 Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.888230 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-httpd" containerID="cri-o://cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783" gracePeriod=30 Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.980523 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9375db1c-410a-439c-bcf0-ede89d471abf-horizon-secret-key\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.980625 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-scripts\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.980826 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9375db1c-410a-439c-bcf0-ede89d471abf-logs\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981013 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-config-data\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981078 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f12d17af-6af8-402d-86a3-65079d25114b-horizon-secret-key\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981112 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-config-data\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981154 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12d17af-6af8-402d-86a3-65079d25114b-logs\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981198 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-scripts\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981267 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5xtl\" (UniqueName: \"kubernetes.io/projected/9375db1c-410a-439c-bcf0-ede89d471abf-kube-api-access-h5xtl\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.981353 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p4m4\" (UniqueName: \"kubernetes.io/projected/f12d17af-6af8-402d-86a3-65079d25114b-kube-api-access-2p4m4\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.982890 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-config-data\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.983719 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12d17af-6af8-402d-86a3-65079d25114b-logs\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:12 crc kubenswrapper[4758]: I1203 18:41:12.984281 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-scripts\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.002447 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f12d17af-6af8-402d-86a3-65079d25114b-horizon-secret-key\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.005968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p4m4\" (UniqueName: \"kubernetes.io/projected/f12d17af-6af8-402d-86a3-65079d25114b-kube-api-access-2p4m4\") pod \"horizon-7746fd787c-t4ld6\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.038475 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.083987 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5xtl\" (UniqueName: \"kubernetes.io/projected/9375db1c-410a-439c-bcf0-ede89d471abf-kube-api-access-h5xtl\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.084089 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9375db1c-410a-439c-bcf0-ede89d471abf-horizon-secret-key\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.084119 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-scripts\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.084161 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9375db1c-410a-439c-bcf0-ede89d471abf-logs\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.084252 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-config-data\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.088163 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9375db1c-410a-439c-bcf0-ede89d471abf-horizon-secret-key\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.088675 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9375db1c-410a-439c-bcf0-ede89d471abf-logs\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.088996 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-scripts\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.089359 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-config-data\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.103356 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5xtl\" (UniqueName: \"kubernetes.io/projected/9375db1c-410a-439c-bcf0-ede89d471abf-kube-api-access-h5xtl\") pod \"horizon-74c78b48d9-82vjv\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.150415 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.556151 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7746fd787c-t4ld6"] Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.648433 4758 generic.go:334] "Generic (PLEG): container finished" podID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerID="303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38" exitCode=143 Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.648457 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f386b68a-b4be-459f-a332-f93ec438ddcb","Type":"ContainerDied","Data":"303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38"} Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.651043 4758 generic.go:334] "Generic (PLEG): container finished" podID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerID="17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98" exitCode=143 Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.651256 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e611edbe-5bdd-4061-baaf-fc92217e4eaf","Type":"ContainerDied","Data":"17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98"} Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.652610 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7746fd787c-t4ld6" event={"ID":"f12d17af-6af8-402d-86a3-65079d25114b","Type":"ContainerStarted","Data":"cfa6bed1bd21b17d122111016e5948ca56fa05e8c0721f1f675bdb342063511d"} Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.689613 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74c78b48d9-82vjv"] Dec 03 18:41:13 crc kubenswrapper[4758]: W1203 18:41:13.692866 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9375db1c_410a_439c_bcf0_ede89d471abf.slice/crio-82de84ccd3923d9e08dc828c9a1185b6106afea9445f3319e759396729b8a04b WatchSource:0}: Error finding container 82de84ccd3923d9e08dc828c9a1185b6106afea9445f3319e759396729b8a04b: Status 404 returned error can't find the container with id 82de84ccd3923d9e08dc828c9a1185b6106afea9445f3319e759396729b8a04b Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.726434 4758 scope.go:117] "RemoveContainer" containerID="e4b7f12e3376095d8a766754ee6f69323f464a37d6ba96f406d01a63d66f8ecc" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.765905 4758 scope.go:117] "RemoveContainer" containerID="4ad5b88232968457a3211d702994039837d6b4ce2209d4b6e178c85feb85cd26" Dec 03 18:41:13 crc kubenswrapper[4758]: I1203 18:41:13.795857 4758 scope.go:117] "RemoveContainer" containerID="ce44fe8deaa6b02a4616b2d7a016824c7a4b59f50dbebe2495485b924d55501e" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:13.999509 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74c78b48d9-82vjv"] Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.038131 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-df5696bb9-ckrnz"] Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.042141 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.054412 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-df5696bb9-ckrnz"] Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.206218 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339a325b-d675-41c1-87cb-22b328569743-logs\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.206604 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/339a325b-d675-41c1-87cb-22b328569743-horizon-secret-key\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.206642 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4vw6\" (UniqueName: \"kubernetes.io/projected/339a325b-d675-41c1-87cb-22b328569743-kube-api-access-f4vw6\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.206871 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-config-data\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.206954 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-scripts\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.309629 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339a325b-d675-41c1-87cb-22b328569743-logs\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.309821 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/339a325b-d675-41c1-87cb-22b328569743-horizon-secret-key\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.309851 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4vw6\" (UniqueName: \"kubernetes.io/projected/339a325b-d675-41c1-87cb-22b328569743-kube-api-access-f4vw6\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.309905 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-config-data\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.309996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-scripts\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.310200 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339a325b-d675-41c1-87cb-22b328569743-logs\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.310883 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-scripts\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.311263 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-config-data\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.316594 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/339a325b-d675-41c1-87cb-22b328569743-horizon-secret-key\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.327531 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4vw6\" (UniqueName: \"kubernetes.io/projected/339a325b-d675-41c1-87cb-22b328569743-kube-api-access-f4vw6\") pod \"horizon-df5696bb9-ckrnz\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.360110 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.704203 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c78b48d9-82vjv" event={"ID":"9375db1c-410a-439c-bcf0-ede89d471abf","Type":"ContainerStarted","Data":"82de84ccd3923d9e08dc828c9a1185b6106afea9445f3319e759396729b8a04b"} Dec 03 18:41:14 crc kubenswrapper[4758]: I1203 18:41:14.880344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-df5696bb9-ckrnz"] Dec 03 18:41:14 crc kubenswrapper[4758]: W1203 18:41:14.909076 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod339a325b_d675_41c1_87cb_22b328569743.slice/crio-42c82137443eaad9b418e57b73c2d6ef6e45d60309dd37079e872e995b2a11c8 WatchSource:0}: Error finding container 42c82137443eaad9b418e57b73c2d6ef6e45d60309dd37079e872e995b2a11c8: Status 404 returned error can't find the container with id 42c82137443eaad9b418e57b73c2d6ef6e45d60309dd37079e872e995b2a11c8 Dec 03 18:41:15 crc kubenswrapper[4758]: I1203 18:41:15.717525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-df5696bb9-ckrnz" event={"ID":"339a325b-d675-41c1-87cb-22b328569743","Type":"ContainerStarted","Data":"42c82137443eaad9b418e57b73c2d6ef6e45d60309dd37079e872e995b2a11c8"} Dec 03 18:41:15 crc kubenswrapper[4758]: I1203 18:41:15.908823 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.1.46:9292/healthcheck\": read tcp 10.217.0.2:46400->10.217.1.46:9292: read: connection reset by peer" Dec 03 18:41:15 crc kubenswrapper[4758]: I1203 18:41:15.908919 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.1.46:9292/healthcheck\": read tcp 10.217.0.2:46410->10.217.1.46:9292: read: connection reset by peer" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.556941 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.662899 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-config-data\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.663009 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-combined-ca-bundle\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.663116 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xrd4\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-kube-api-access-2xrd4\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.663155 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-ceph\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.663198 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-logs\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.663377 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-httpd-run\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.663429 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-scripts\") pod \"f386b68a-b4be-459f-a332-f93ec438ddcb\" (UID: \"f386b68a-b4be-459f-a332-f93ec438ddcb\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.665514 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-logs" (OuterVolumeSpecName: "logs") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.665641 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.676719 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-scripts" (OuterVolumeSpecName: "scripts") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.681441 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-ceph" (OuterVolumeSpecName: "ceph") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.681497 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-kube-api-access-2xrd4" (OuterVolumeSpecName: "kube-api-access-2xrd4") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "kube-api-access-2xrd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.711155 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.769384 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.769426 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xrd4\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-kube-api-access-2xrd4\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.769442 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f386b68a-b4be-459f-a332-f93ec438ddcb-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.769459 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.769470 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f386b68a-b4be-459f-a332-f93ec438ddcb-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.781527 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.790935 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.792209 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f386b68a-b4be-459f-a332-f93ec438ddcb","Type":"ContainerDied","Data":"0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179"} Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.792439 4758 scope.go:117] "RemoveContainer" containerID="0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.793134 4758 generic.go:334] "Generic (PLEG): container finished" podID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerID="0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179" exitCode=0 Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.793284 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f386b68a-b4be-459f-a332-f93ec438ddcb","Type":"ContainerDied","Data":"8231711a22fe55f0bc62b377117f1b3579ac340571e3c177c27febc049134a3a"} Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.808890 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-config-data" (OuterVolumeSpecName: "config-data") pod "f386b68a-b4be-459f-a332-f93ec438ddcb" (UID: "f386b68a-b4be-459f-a332-f93ec438ddcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.817622 4758 generic.go:334] "Generic (PLEG): container finished" podID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerID="cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783" exitCode=0 Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.817714 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.817716 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e611edbe-5bdd-4061-baaf-fc92217e4eaf","Type":"ContainerDied","Data":"cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783"} Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.817885 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e611edbe-5bdd-4061-baaf-fc92217e4eaf","Type":"ContainerDied","Data":"ff7736751c3a239f6d246c17c997ba3ddde2283ad1194218bc8b0237c4da8b29"} Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872532 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-ceph\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872605 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-combined-ca-bundle\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872724 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-httpd-run\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872762 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-config-data\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872784 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx2cj\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-kube-api-access-hx2cj\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872864 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-scripts\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.872967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-logs\") pod \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\" (UID: \"e611edbe-5bdd-4061-baaf-fc92217e4eaf\") " Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.876552 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.876590 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f386b68a-b4be-459f-a332-f93ec438ddcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.877227 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-logs" (OuterVolumeSpecName: "logs") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.880439 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-ceph" (OuterVolumeSpecName: "ceph") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.883817 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.884275 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-kube-api-access-hx2cj" (OuterVolumeSpecName: "kube-api-access-hx2cj") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "kube-api-access-hx2cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.885652 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-scripts" (OuterVolumeSpecName: "scripts") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.910280 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.935269 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-config-data" (OuterVolumeSpecName: "config-data") pod "e611edbe-5bdd-4061-baaf-fc92217e4eaf" (UID: "e611edbe-5bdd-4061-baaf-fc92217e4eaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979607 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979732 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979751 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979764 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979775 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx2cj\" (UniqueName: \"kubernetes.io/projected/e611edbe-5bdd-4061-baaf-fc92217e4eaf-kube-api-access-hx2cj\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979787 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e611edbe-5bdd-4061-baaf-fc92217e4eaf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:16 crc kubenswrapper[4758]: I1203 18:41:16.979797 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e611edbe-5bdd-4061-baaf-fc92217e4eaf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.152044 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.166375 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183276 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: E1203 18:41:17.183651 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-log" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183666 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-log" Dec 03 18:41:17 crc kubenswrapper[4758]: E1203 18:41:17.183695 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-log" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183703 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-log" Dec 03 18:41:17 crc kubenswrapper[4758]: E1203 18:41:17.183715 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-httpd" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183721 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-httpd" Dec 03 18:41:17 crc kubenswrapper[4758]: E1203 18:41:17.183748 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-httpd" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183754 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-httpd" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183933 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-log" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183951 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-httpd" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183967 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" containerName="glance-httpd" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.183977 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" containerName="glance-log" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.184972 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.188138 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cx5sq" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.188376 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.188505 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.209484 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.227818 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.238008 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.250396 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.252543 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.260288 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.260921 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287001 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287076 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-config-data\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287108 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-logs\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287223 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdthb\" (UniqueName: \"kubernetes.io/projected/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-kube-api-access-cdthb\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287286 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287507 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-ceph\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.287582 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-scripts\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.392695 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/579d0799-97e3-468a-afcb-6c0ca7c67004-logs\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.392750 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/579d0799-97e3-468a-afcb-6c0ca7c67004-ceph\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.392865 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-ceph\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.392956 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-config-data\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393032 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-scripts\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393110 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6mz\" (UniqueName: \"kubernetes.io/projected/579d0799-97e3-468a-afcb-6c0ca7c67004-kube-api-access-cg6mz\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393166 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393217 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393244 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-config-data\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393293 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-logs\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393356 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/579d0799-97e3-468a-afcb-6c0ca7c67004-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393379 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdthb\" (UniqueName: \"kubernetes.io/projected/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-kube-api-access-cdthb\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393423 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.393462 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-scripts\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.394522 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.402876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-logs\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.421188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-config-data\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.422203 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-ceph\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.429290 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-scripts\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.431586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdthb\" (UniqueName: \"kubernetes.io/projected/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-kube-api-access-cdthb\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.432241 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aca4a33e-8dac-4d3b-bd2b-7ef196a94871-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aca4a33e-8dac-4d3b-bd2b-7ef196a94871\") " pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495177 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/579d0799-97e3-468a-afcb-6c0ca7c67004-logs\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495220 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/579d0799-97e3-468a-afcb-6c0ca7c67004-ceph\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495287 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-config-data\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495327 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6mz\" (UniqueName: \"kubernetes.io/projected/579d0799-97e3-468a-afcb-6c0ca7c67004-kube-api-access-cg6mz\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495468 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/579d0799-97e3-468a-afcb-6c0ca7c67004-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.495496 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-scripts\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.497955 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/579d0799-97e3-468a-afcb-6c0ca7c67004-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.498182 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/579d0799-97e3-468a-afcb-6c0ca7c67004-logs\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.499218 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-scripts\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.500662 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-config-data\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.510163 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/579d0799-97e3-468a-afcb-6c0ca7c67004-ceph\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.510576 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579d0799-97e3-468a-afcb-6c0ca7c67004-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.517304 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.517398 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6mz\" (UniqueName: \"kubernetes.io/projected/579d0799-97e3-468a-afcb-6c0ca7c67004-kube-api-access-cg6mz\") pod \"glance-default-internal-api-0\" (UID: \"579d0799-97e3-468a-afcb-6c0ca7c67004\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:41:17 crc kubenswrapper[4758]: I1203 18:41:17.574887 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:19 crc kubenswrapper[4758]: I1203 18:41:19.127325 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e611edbe-5bdd-4061-baaf-fc92217e4eaf" path="/var/lib/kubelet/pods/e611edbe-5bdd-4061-baaf-fc92217e4eaf/volumes" Dec 03 18:41:19 crc kubenswrapper[4758]: I1203 18:41:19.128221 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f386b68a-b4be-459f-a332-f93ec438ddcb" path="/var/lib/kubelet/pods/f386b68a-b4be-459f-a332-f93ec438ddcb/volumes" Dec 03 18:41:21 crc kubenswrapper[4758]: I1203 18:41:21.986176 4758 scope.go:117] "RemoveContainer" containerID="303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.419053 4758 scope.go:117] "RemoveContainer" containerID="0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179" Dec 03 18:41:22 crc kubenswrapper[4758]: E1203 18:41:22.420790 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179\": container with ID starting with 0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179 not found: ID does not exist" containerID="0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.420844 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179"} err="failed to get container status \"0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179\": rpc error: code = NotFound desc = could not find container \"0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179\": container with ID starting with 0b1c1538f5deeac85461e6df72386db2d20fcf4b56e5133d155f2ff9c27a6179 not found: ID does not exist" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.420877 4758 scope.go:117] "RemoveContainer" containerID="303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38" Dec 03 18:41:22 crc kubenswrapper[4758]: E1203 18:41:22.421210 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38\": container with ID starting with 303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38 not found: ID does not exist" containerID="303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.421248 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38"} err="failed to get container status \"303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38\": rpc error: code = NotFound desc = could not find container \"303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38\": container with ID starting with 303f7d4af867b270de2f016ccce0f5c9502c054f546d6c9bad91e1bf6fd1cd38 not found: ID does not exist" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.421275 4758 scope.go:117] "RemoveContainer" containerID="cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.466180 4758 scope.go:117] "RemoveContainer" containerID="17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.507137 4758 scope.go:117] "RemoveContainer" containerID="cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783" Dec 03 18:41:22 crc kubenswrapper[4758]: E1203 18:41:22.507803 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783\": container with ID starting with cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783 not found: ID does not exist" containerID="cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.507842 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783"} err="failed to get container status \"cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783\": rpc error: code = NotFound desc = could not find container \"cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783\": container with ID starting with cce41b438655eab43ef7d0baeef543dfe55334fec28eead73683fb7a6655e783 not found: ID does not exist" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.507868 4758 scope.go:117] "RemoveContainer" containerID="17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98" Dec 03 18:41:22 crc kubenswrapper[4758]: E1203 18:41:22.508289 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98\": container with ID starting with 17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98 not found: ID does not exist" containerID="17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.508315 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98"} err="failed to get container status \"17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98\": rpc error: code = NotFound desc = could not find container \"17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98\": container with ID starting with 17bc77759b1083172394634a9e4d442ac71d480005d6ef809f780e6be3801e98 not found: ID does not exist" Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.630072 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.857901 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.896216 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-df5696bb9-ckrnz" event={"ID":"339a325b-d675-41c1-87cb-22b328569743","Type":"ContainerStarted","Data":"ad22110c038980687e63440cd8ad0b1244374f2f6d96881a371a94c6993293b8"} Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.903530 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c78b48d9-82vjv" event={"ID":"9375db1c-410a-439c-bcf0-ede89d471abf","Type":"ContainerStarted","Data":"7412f585940e1590344ccaa4e9a9d6f736a540717529cbaf5945b46388574fbb"} Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.905003 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"579d0799-97e3-468a-afcb-6c0ca7c67004","Type":"ContainerStarted","Data":"72566ae0d07f2128e6142fb9dc42a3cc14bf8ae798947dbda3339461d7ad61ab"} Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.908453 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7746fd787c-t4ld6" event={"ID":"f12d17af-6af8-402d-86a3-65079d25114b","Type":"ContainerStarted","Data":"c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d"} Dec 03 18:41:22 crc kubenswrapper[4758]: I1203 18:41:22.911284 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aca4a33e-8dac-4d3b-bd2b-7ef196a94871","Type":"ContainerStarted","Data":"54304c9dacdfe6935a10105e37770d3838f985c00f590e5cc7cdff0ab1d6b4af"} Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.931557 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c78b48d9-82vjv" event={"ID":"9375db1c-410a-439c-bcf0-ede89d471abf","Type":"ContainerStarted","Data":"2850d6ac7f4fcf029b36d31c28e133c3c6e1c57a50afaf1e0b39293f115e1c33"} Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.931647 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74c78b48d9-82vjv" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon-log" containerID="cri-o://7412f585940e1590344ccaa4e9a9d6f736a540717529cbaf5945b46388574fbb" gracePeriod=30 Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.931742 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74c78b48d9-82vjv" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon" containerID="cri-o://2850d6ac7f4fcf029b36d31c28e133c3c6e1c57a50afaf1e0b39293f115e1c33" gracePeriod=30 Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.935882 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"579d0799-97e3-468a-afcb-6c0ca7c67004","Type":"ContainerStarted","Data":"14ecd80e08b6c497d1ba7663aa5f674a7210f0a4eb4a5684d4a57f4ecd30224a"} Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.935955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"579d0799-97e3-468a-afcb-6c0ca7c67004","Type":"ContainerStarted","Data":"cef47cc78f08c12082c8d8c5f65e919a686ac167515cd2cf90b75f070ae6c595"} Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.947377 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7746fd787c-t4ld6" event={"ID":"f12d17af-6af8-402d-86a3-65079d25114b","Type":"ContainerStarted","Data":"b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb"} Dec 03 18:41:23 crc kubenswrapper[4758]: I1203 18:41:23.950118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aca4a33e-8dac-4d3b-bd2b-7ef196a94871","Type":"ContainerStarted","Data":"7e355eaab108d6a6af560efe69bc1096c10a5a533d0f47e7e95f8ed95e0426f9"} Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.015959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-df5696bb9-ckrnz" event={"ID":"339a325b-d675-41c1-87cb-22b328569743","Type":"ContainerStarted","Data":"24776dc59fd5b45da37bd7cd1d0a8645987082b054e38c439bb4a3b272dd280e"} Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.084784 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7746fd787c-t4ld6" podStartSLOduration=3.228687356 podStartE2EDuration="12.084759943s" podCreationTimestamp="2025-12-03 18:41:12 +0000 UTC" firstStartedPulling="2025-12-03 18:41:13.568519584 +0000 UTC m=+6328.769896445" lastFinishedPulling="2025-12-03 18:41:22.424592171 +0000 UTC m=+6337.625969032" observedRunningTime="2025-12-03 18:41:24.041620434 +0000 UTC m=+6339.242997295" watchObservedRunningTime="2025-12-03 18:41:24.084759943 +0000 UTC m=+6339.286136814" Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.104151 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74c78b48d9-82vjv" podStartSLOduration=3.330038152 podStartE2EDuration="12.104131958s" podCreationTimestamp="2025-12-03 18:41:12 +0000 UTC" firstStartedPulling="2025-12-03 18:41:13.695410861 +0000 UTC m=+6328.896787722" lastFinishedPulling="2025-12-03 18:41:22.469504667 +0000 UTC m=+6337.670881528" observedRunningTime="2025-12-03 18:41:23.975289118 +0000 UTC m=+6339.176665999" watchObservedRunningTime="2025-12-03 18:41:24.104131958 +0000 UTC m=+6339.305508819" Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.124642 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.124621722 podStartE2EDuration="7.124621722s" podCreationTimestamp="2025-12-03 18:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:41:24.110405087 +0000 UTC m=+6339.311781958" watchObservedRunningTime="2025-12-03 18:41:24.124621722 +0000 UTC m=+6339.325998583" Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.216268 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-df5696bb9-ckrnz" podStartSLOduration=2.703358774 podStartE2EDuration="10.216244984s" podCreationTimestamp="2025-12-03 18:41:14 +0000 UTC" firstStartedPulling="2025-12-03 18:41:14.912348449 +0000 UTC m=+6330.113725310" lastFinishedPulling="2025-12-03 18:41:22.425234659 +0000 UTC m=+6337.626611520" observedRunningTime="2025-12-03 18:41:24.196340314 +0000 UTC m=+6339.397717175" watchObservedRunningTime="2025-12-03 18:41:24.216244984 +0000 UTC m=+6339.417621845" Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.361041 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:24 crc kubenswrapper[4758]: I1203 18:41:24.361107 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:25 crc kubenswrapper[4758]: I1203 18:41:25.026146 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aca4a33e-8dac-4d3b-bd2b-7ef196a94871","Type":"ContainerStarted","Data":"e0e1c2e3c6d733c798e94fbd2b4a1a584c37c5d004f4d39e962da9569177af95"} Dec 03 18:41:25 crc kubenswrapper[4758]: I1203 18:41:25.057573 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.057551659 podStartE2EDuration="8.057551659s" podCreationTimestamp="2025-12-03 18:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:41:25.042210933 +0000 UTC m=+6340.243587804" watchObservedRunningTime="2025-12-03 18:41:25.057551659 +0000 UTC m=+6340.258928520" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.517751 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.518378 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.548573 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.558882 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.575894 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.575942 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.623160 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:27 crc kubenswrapper[4758]: I1203 18:41:27.628660 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:28 crc kubenswrapper[4758]: I1203 18:41:28.055395 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 18:41:28 crc kubenswrapper[4758]: I1203 18:41:28.055699 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:28 crc kubenswrapper[4758]: I1203 18:41:28.055771 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:28 crc kubenswrapper[4758]: I1203 18:41:28.055859 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 18:41:29 crc kubenswrapper[4758]: I1203 18:41:29.053212 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9hgcn"] Dec 03 18:41:29 crc kubenswrapper[4758]: I1203 18:41:29.063201 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b074-account-create-update-j4m2k"] Dec 03 18:41:29 crc kubenswrapper[4758]: I1203 18:41:29.072976 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b074-account-create-update-j4m2k"] Dec 03 18:41:29 crc kubenswrapper[4758]: I1203 18:41:29.084055 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9hgcn"] Dec 03 18:41:29 crc kubenswrapper[4758]: I1203 18:41:29.127345 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbb51fa-cc6e-4d84-a488-c67712eeb4f3" path="/var/lib/kubelet/pods/2dbb51fa-cc6e-4d84-a488-c67712eeb4f3/volumes" Dec 03 18:41:29 crc kubenswrapper[4758]: I1203 18:41:29.128193 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c9563ce-efff-453a-8c17-0704001669ba" path="/var/lib/kubelet/pods/8c9563ce-efff-453a-8c17-0704001669ba/volumes" Dec 03 18:41:30 crc kubenswrapper[4758]: I1203 18:41:30.796748 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 18:41:30 crc kubenswrapper[4758]: I1203 18:41:30.982605 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:30 crc kubenswrapper[4758]: I1203 18:41:30.982697 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 18:41:30 crc kubenswrapper[4758]: I1203 18:41:30.983730 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 18:41:31 crc kubenswrapper[4758]: I1203 18:41:31.512481 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 18:41:33 crc kubenswrapper[4758]: I1203 18:41:33.039832 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:33 crc kubenswrapper[4758]: I1203 18:41:33.041312 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:33 crc kubenswrapper[4758]: I1203 18:41:33.042321 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7746fd787c-t4ld6" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 03 18:41:33 crc kubenswrapper[4758]: I1203 18:41:33.151843 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:34 crc kubenswrapper[4758]: I1203 18:41:34.362175 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-df5696bb9-ckrnz" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 03 18:41:39 crc kubenswrapper[4758]: I1203 18:41:39.039361 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-k4w98"] Dec 03 18:41:39 crc kubenswrapper[4758]: I1203 18:41:39.056009 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-k4w98"] Dec 03 18:41:39 crc kubenswrapper[4758]: I1203 18:41:39.130263 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91dd638a-93de-4683-9e4a-115bb621cd38" path="/var/lib/kubelet/pods/91dd638a-93de-4683-9e4a-115bb621cd38/volumes" Dec 03 18:41:44 crc kubenswrapper[4758]: I1203 18:41:44.961200 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:46 crc kubenswrapper[4758]: I1203 18:41:46.178309 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:46 crc kubenswrapper[4758]: I1203 18:41:46.643455 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:41:47 crc kubenswrapper[4758]: I1203 18:41:47.896245 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:41:47 crc kubenswrapper[4758]: I1203 18:41:47.989386 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7746fd787c-t4ld6"] Dec 03 18:41:47 crc kubenswrapper[4758]: I1203 18:41:47.989633 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7746fd787c-t4ld6" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon-log" containerID="cri-o://c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d" gracePeriod=30 Dec 03 18:41:47 crc kubenswrapper[4758]: I1203 18:41:47.989720 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7746fd787c-t4ld6" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" containerID="cri-o://b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb" gracePeriod=30 Dec 03 18:41:51 crc kubenswrapper[4758]: I1203 18:41:51.320753 4758 generic.go:334] "Generic (PLEG): container finished" podID="f12d17af-6af8-402d-86a3-65079d25114b" containerID="b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb" exitCode=0 Dec 03 18:41:51 crc kubenswrapper[4758]: I1203 18:41:51.320866 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7746fd787c-t4ld6" event={"ID":"f12d17af-6af8-402d-86a3-65079d25114b","Type":"ContainerDied","Data":"b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb"} Dec 03 18:41:53 crc kubenswrapper[4758]: I1203 18:41:53.039384 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7746fd787c-t4ld6" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.353757 4758 generic.go:334] "Generic (PLEG): container finished" podID="9375db1c-410a-439c-bcf0-ede89d471abf" containerID="2850d6ac7f4fcf029b36d31c28e133c3c6e1c57a50afaf1e0b39293f115e1c33" exitCode=137 Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.354110 4758 generic.go:334] "Generic (PLEG): container finished" podID="9375db1c-410a-439c-bcf0-ede89d471abf" containerID="7412f585940e1590344ccaa4e9a9d6f736a540717529cbaf5945b46388574fbb" exitCode=137 Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.353799 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c78b48d9-82vjv" event={"ID":"9375db1c-410a-439c-bcf0-ede89d471abf","Type":"ContainerDied","Data":"2850d6ac7f4fcf029b36d31c28e133c3c6e1c57a50afaf1e0b39293f115e1c33"} Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.354142 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c78b48d9-82vjv" event={"ID":"9375db1c-410a-439c-bcf0-ede89d471abf","Type":"ContainerDied","Data":"7412f585940e1590344ccaa4e9a9d6f736a540717529cbaf5945b46388574fbb"} Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.354153 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74c78b48d9-82vjv" event={"ID":"9375db1c-410a-439c-bcf0-ede89d471abf","Type":"ContainerDied","Data":"82de84ccd3923d9e08dc828c9a1185b6106afea9445f3319e759396729b8a04b"} Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.354163 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82de84ccd3923d9e08dc828c9a1185b6106afea9445f3319e759396729b8a04b" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.419694 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.475175 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-config-data\") pod \"9375db1c-410a-439c-bcf0-ede89d471abf\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.502128 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-config-data" (OuterVolumeSpecName: "config-data") pod "9375db1c-410a-439c-bcf0-ede89d471abf" (UID: "9375db1c-410a-439c-bcf0-ede89d471abf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.576745 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5xtl\" (UniqueName: \"kubernetes.io/projected/9375db1c-410a-439c-bcf0-ede89d471abf-kube-api-access-h5xtl\") pod \"9375db1c-410a-439c-bcf0-ede89d471abf\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.577048 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9375db1c-410a-439c-bcf0-ede89d471abf-horizon-secret-key\") pod \"9375db1c-410a-439c-bcf0-ede89d471abf\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.577220 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-scripts\") pod \"9375db1c-410a-439c-bcf0-ede89d471abf\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.577338 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9375db1c-410a-439c-bcf0-ede89d471abf-logs\") pod \"9375db1c-410a-439c-bcf0-ede89d471abf\" (UID: \"9375db1c-410a-439c-bcf0-ede89d471abf\") " Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.577836 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9375db1c-410a-439c-bcf0-ede89d471abf-logs" (OuterVolumeSpecName: "logs") pod "9375db1c-410a-439c-bcf0-ede89d471abf" (UID: "9375db1c-410a-439c-bcf0-ede89d471abf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.578046 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.580475 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9375db1c-410a-439c-bcf0-ede89d471abf-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9375db1c-410a-439c-bcf0-ede89d471abf" (UID: "9375db1c-410a-439c-bcf0-ede89d471abf"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.580515 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9375db1c-410a-439c-bcf0-ede89d471abf-kube-api-access-h5xtl" (OuterVolumeSpecName: "kube-api-access-h5xtl") pod "9375db1c-410a-439c-bcf0-ede89d471abf" (UID: "9375db1c-410a-439c-bcf0-ede89d471abf"). InnerVolumeSpecName "kube-api-access-h5xtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.618614 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-scripts" (OuterVolumeSpecName: "scripts") pod "9375db1c-410a-439c-bcf0-ede89d471abf" (UID: "9375db1c-410a-439c-bcf0-ede89d471abf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.679124 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5xtl\" (UniqueName: \"kubernetes.io/projected/9375db1c-410a-439c-bcf0-ede89d471abf-kube-api-access-h5xtl\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.679161 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9375db1c-410a-439c-bcf0-ede89d471abf-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.679171 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9375db1c-410a-439c-bcf0-ede89d471abf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:54 crc kubenswrapper[4758]: I1203 18:41:54.679179 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9375db1c-410a-439c-bcf0-ede89d471abf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:41:55 crc kubenswrapper[4758]: I1203 18:41:55.369397 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74c78b48d9-82vjv" Dec 03 18:41:55 crc kubenswrapper[4758]: I1203 18:41:55.402890 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74c78b48d9-82vjv"] Dec 03 18:41:55 crc kubenswrapper[4758]: I1203 18:41:55.413412 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74c78b48d9-82vjv"] Dec 03 18:41:57 crc kubenswrapper[4758]: I1203 18:41:57.128421 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" path="/var/lib/kubelet/pods/9375db1c-410a-439c-bcf0-ede89d471abf/volumes" Dec 03 18:42:03 crc kubenswrapper[4758]: I1203 18:42:03.039653 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7746fd787c-t4ld6" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 03 18:42:11 crc kubenswrapper[4758]: I1203 18:42:11.050620 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-bd30-account-create-update-r7bsr"] Dec 03 18:42:11 crc kubenswrapper[4758]: I1203 18:42:11.061556 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-shwqh"] Dec 03 18:42:11 crc kubenswrapper[4758]: I1203 18:42:11.073481 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-shwqh"] Dec 03 18:42:11 crc kubenswrapper[4758]: I1203 18:42:11.083042 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-bd30-account-create-update-r7bsr"] Dec 03 18:42:11 crc kubenswrapper[4758]: I1203 18:42:11.125149 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52fdf77e-873b-47f1-9c59-6f46226c43a6" path="/var/lib/kubelet/pods/52fdf77e-873b-47f1-9c59-6f46226c43a6/volumes" Dec 03 18:42:11 crc kubenswrapper[4758]: I1203 18:42:11.125826 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab563bad-7d6e-4c2e-bccd-b882230dcc9c" path="/var/lib/kubelet/pods/ab563bad-7d6e-4c2e-bccd-b882230dcc9c/volumes" Dec 03 18:42:13 crc kubenswrapper[4758]: I1203 18:42:13.039311 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7746fd787c-t4ld6" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 03 18:42:13 crc kubenswrapper[4758]: I1203 18:42:13.039765 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:42:13 crc kubenswrapper[4758]: I1203 18:42:13.875472 4758 scope.go:117] "RemoveContainer" containerID="5bd75ec17d038ee42069ba402b6f27348155bd2da28cb6a2cc402dd65fdb3df6" Dec 03 18:42:13 crc kubenswrapper[4758]: I1203 18:42:13.916786 4758 scope.go:117] "RemoveContainer" containerID="b5532523573bc468a8cc42820ae56e455235eb52b6ee507c2ab8308b02ccc779" Dec 03 18:42:13 crc kubenswrapper[4758]: I1203 18:42:13.989917 4758 scope.go:117] "RemoveContainer" containerID="c3220aded8dc6ef27ba410012855f7bc774d86bb755f01b192172bf67754e6d8" Dec 03 18:42:14 crc kubenswrapper[4758]: I1203 18:42:14.042666 4758 scope.go:117] "RemoveContainer" containerID="2f114b2c763b993886f79ab64e17c1978c8f3fff8ad14e435968b492daa39e76" Dec 03 18:42:14 crc kubenswrapper[4758]: I1203 18:42:14.084174 4758 scope.go:117] "RemoveContainer" containerID="467e787a7a3dafe0bb1f9481f0a73f819ed5bfdc06681a73212291467ff1628e" Dec 03 18:42:17 crc kubenswrapper[4758]: I1203 18:42:17.061439 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dkgcn"] Dec 03 18:42:17 crc kubenswrapper[4758]: I1203 18:42:17.084117 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dkgcn"] Dec 03 18:42:17 crc kubenswrapper[4758]: I1203 18:42:17.129876 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0d1b1d-8465-47f6-89ad-a05266e6acaf" path="/var/lib/kubelet/pods/6e0d1b1d-8465-47f6-89ad-a05266e6acaf/volumes" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.429531 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.619526 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p4m4\" (UniqueName: \"kubernetes.io/projected/f12d17af-6af8-402d-86a3-65079d25114b-kube-api-access-2p4m4\") pod \"f12d17af-6af8-402d-86a3-65079d25114b\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.619883 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12d17af-6af8-402d-86a3-65079d25114b-logs\") pod \"f12d17af-6af8-402d-86a3-65079d25114b\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.620008 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f12d17af-6af8-402d-86a3-65079d25114b-horizon-secret-key\") pod \"f12d17af-6af8-402d-86a3-65079d25114b\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.620130 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-scripts\") pod \"f12d17af-6af8-402d-86a3-65079d25114b\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.620291 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-config-data\") pod \"f12d17af-6af8-402d-86a3-65079d25114b\" (UID: \"f12d17af-6af8-402d-86a3-65079d25114b\") " Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.620390 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f12d17af-6af8-402d-86a3-65079d25114b-logs" (OuterVolumeSpecName: "logs") pod "f12d17af-6af8-402d-86a3-65079d25114b" (UID: "f12d17af-6af8-402d-86a3-65079d25114b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.621284 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12d17af-6af8-402d-86a3-65079d25114b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.625208 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f12d17af-6af8-402d-86a3-65079d25114b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f12d17af-6af8-402d-86a3-65079d25114b" (UID: "f12d17af-6af8-402d-86a3-65079d25114b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.625900 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12d17af-6af8-402d-86a3-65079d25114b-kube-api-access-2p4m4" (OuterVolumeSpecName: "kube-api-access-2p4m4") pod "f12d17af-6af8-402d-86a3-65079d25114b" (UID: "f12d17af-6af8-402d-86a3-65079d25114b"). InnerVolumeSpecName "kube-api-access-2p4m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.655261 4758 generic.go:334] "Generic (PLEG): container finished" podID="f12d17af-6af8-402d-86a3-65079d25114b" containerID="c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d" exitCode=137 Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.655332 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7746fd787c-t4ld6" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.655335 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7746fd787c-t4ld6" event={"ID":"f12d17af-6af8-402d-86a3-65079d25114b","Type":"ContainerDied","Data":"c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d"} Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.655591 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7746fd787c-t4ld6" event={"ID":"f12d17af-6af8-402d-86a3-65079d25114b","Type":"ContainerDied","Data":"cfa6bed1bd21b17d122111016e5948ca56fa05e8c0721f1f675bdb342063511d"} Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.655626 4758 scope.go:117] "RemoveContainer" containerID="b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.660634 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-config-data" (OuterVolumeSpecName: "config-data") pod "f12d17af-6af8-402d-86a3-65079d25114b" (UID: "f12d17af-6af8-402d-86a3-65079d25114b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.662650 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-scripts" (OuterVolumeSpecName: "scripts") pod "f12d17af-6af8-402d-86a3-65079d25114b" (UID: "f12d17af-6af8-402d-86a3-65079d25114b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.723138 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.723177 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p4m4\" (UniqueName: \"kubernetes.io/projected/f12d17af-6af8-402d-86a3-65079d25114b-kube-api-access-2p4m4\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.723189 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f12d17af-6af8-402d-86a3-65079d25114b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.723199 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f12d17af-6af8-402d-86a3-65079d25114b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:18 crc kubenswrapper[4758]: I1203 18:42:18.950406 4758 scope.go:117] "RemoveContainer" containerID="c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d" Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.002970 4758 scope.go:117] "RemoveContainer" containerID="b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb" Dec 03 18:42:19 crc kubenswrapper[4758]: E1203 18:42:19.004497 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb\": container with ID starting with b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb not found: ID does not exist" containerID="b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb" Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.004546 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb"} err="failed to get container status \"b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb\": rpc error: code = NotFound desc = could not find container \"b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb\": container with ID starting with b534865e1a12d3ba8bdefa3262e26badfeb5d51a451cf9419383a134a4a0fdeb not found: ID does not exist" Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.004579 4758 scope.go:117] "RemoveContainer" containerID="c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d" Dec 03 18:42:19 crc kubenswrapper[4758]: E1203 18:42:19.005155 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d\": container with ID starting with c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d not found: ID does not exist" containerID="c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d" Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.005233 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d"} err="failed to get container status \"c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d\": rpc error: code = NotFound desc = could not find container \"c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d\": container with ID starting with c0c7005c6f5224bdf67a1172f428dd9c6cd6dfbab21ff88ed47abdce50b1a47d not found: ID does not exist" Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.008551 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7746fd787c-t4ld6"] Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.016783 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7746fd787c-t4ld6"] Dec 03 18:42:19 crc kubenswrapper[4758]: I1203 18:42:19.125538 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12d17af-6af8-402d-86a3-65079d25114b" path="/var/lib/kubelet/pods/f12d17af-6af8-402d-86a3-65079d25114b/volumes" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.516610 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6654448665-7d6rf"] Dec 03 18:42:30 crc kubenswrapper[4758]: E1203 18:42:30.517561 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon-log" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517580 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon-log" Dec 03 18:42:30 crc kubenswrapper[4758]: E1203 18:42:30.517601 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517607 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" Dec 03 18:42:30 crc kubenswrapper[4758]: E1203 18:42:30.517630 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon-log" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517636 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon-log" Dec 03 18:42:30 crc kubenswrapper[4758]: E1203 18:42:30.517647 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517653 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517849 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517872 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon-log" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517883 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9375db1c-410a-439c-bcf0-ede89d471abf" containerName="horizon" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.517897 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12d17af-6af8-402d-86a3-65079d25114b" containerName="horizon-log" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.518989 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.529589 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6654448665-7d6rf"] Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.674483 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/223e253a-8088-4569-935b-7f7bed328067-horizon-secret-key\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.674534 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/223e253a-8088-4569-935b-7f7bed328067-scripts\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.674727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223e253a-8088-4569-935b-7f7bed328067-logs\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.674838 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdm86\" (UniqueName: \"kubernetes.io/projected/223e253a-8088-4569-935b-7f7bed328067-kube-api-access-qdm86\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.674880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/223e253a-8088-4569-935b-7f7bed328067-config-data\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.776341 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/223e253a-8088-4569-935b-7f7bed328067-horizon-secret-key\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.776387 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/223e253a-8088-4569-935b-7f7bed328067-scripts\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.776473 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223e253a-8088-4569-935b-7f7bed328067-logs\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.776555 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdm86\" (UniqueName: \"kubernetes.io/projected/223e253a-8088-4569-935b-7f7bed328067-kube-api-access-qdm86\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.776584 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/223e253a-8088-4569-935b-7f7bed328067-config-data\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.778019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/223e253a-8088-4569-935b-7f7bed328067-scripts\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.778566 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/223e253a-8088-4569-935b-7f7bed328067-config-data\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.778717 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/223e253a-8088-4569-935b-7f7bed328067-logs\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.798413 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/223e253a-8088-4569-935b-7f7bed328067-horizon-secret-key\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.808050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdm86\" (UniqueName: \"kubernetes.io/projected/223e253a-8088-4569-935b-7f7bed328067-kube-api-access-qdm86\") pod \"horizon-6654448665-7d6rf\" (UID: \"223e253a-8088-4569-935b-7f7bed328067\") " pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:30 crc kubenswrapper[4758]: I1203 18:42:30.840031 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.518727 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6654448665-7d6rf"] Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.799508 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6654448665-7d6rf" event={"ID":"223e253a-8088-4569-935b-7f7bed328067","Type":"ContainerStarted","Data":"65b7fc922e1f0b42a69d457994f14e0646fbb83a015cb6910ba9c2e20e7a2f92"} Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.867640 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-6kbkf"] Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.869005 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.877088 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-6kbkf"] Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.890784 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-5c0a-account-create-update-dmf7v"] Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.892106 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.898523 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.919144 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-5c0a-account-create-update-dmf7v"] Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.999279 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-operator-scripts\") pod \"heat-5c0a-account-create-update-dmf7v\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.999395 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-operator-scripts\") pod \"heat-db-create-6kbkf\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.999417 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvq2v\" (UniqueName: \"kubernetes.io/projected/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-kube-api-access-pvq2v\") pod \"heat-db-create-6kbkf\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:31 crc kubenswrapper[4758]: I1203 18:42:31.999496 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fg5c\" (UniqueName: \"kubernetes.io/projected/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-kube-api-access-9fg5c\") pod \"heat-5c0a-account-create-update-dmf7v\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.101656 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fg5c\" (UniqueName: \"kubernetes.io/projected/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-kube-api-access-9fg5c\") pod \"heat-5c0a-account-create-update-dmf7v\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.102211 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-operator-scripts\") pod \"heat-5c0a-account-create-update-dmf7v\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.102358 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-operator-scripts\") pod \"heat-db-create-6kbkf\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.102434 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvq2v\" (UniqueName: \"kubernetes.io/projected/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-kube-api-access-pvq2v\") pod \"heat-db-create-6kbkf\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.103253 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-operator-scripts\") pod \"heat-db-create-6kbkf\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.104981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-operator-scripts\") pod \"heat-5c0a-account-create-update-dmf7v\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.117289 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fg5c\" (UniqueName: \"kubernetes.io/projected/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-kube-api-access-9fg5c\") pod \"heat-5c0a-account-create-update-dmf7v\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.118260 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvq2v\" (UniqueName: \"kubernetes.io/projected/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-kube-api-access-pvq2v\") pod \"heat-db-create-6kbkf\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.190622 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.222380 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:32 crc kubenswrapper[4758]: W1203 18:42:32.682804 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3e4a382_3f17_482b_a2c2_cbdc11dbef9c.slice/crio-de40acca1612f9b201cd6ea209437a16f7894d4e3568935a93fe9074e1fff68b WatchSource:0}: Error finding container de40acca1612f9b201cd6ea209437a16f7894d4e3568935a93fe9074e1fff68b: Status 404 returned error can't find the container with id de40acca1612f9b201cd6ea209437a16f7894d4e3568935a93fe9074e1fff68b Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.683061 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-5c0a-account-create-update-dmf7v"] Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.711951 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-6kbkf"] Dec 03 18:42:32 crc kubenswrapper[4758]: W1203 18:42:32.722162 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c2271b8_acf4_4049_a1cd_93a4a7b5b2c5.slice/crio-4a85a96da66480f8d3b1820febb85132ff07c41e09b84c215e4968d947a0a18f WatchSource:0}: Error finding container 4a85a96da66480f8d3b1820febb85132ff07c41e09b84c215e4968d947a0a18f: Status 404 returned error can't find the container with id 4a85a96da66480f8d3b1820febb85132ff07c41e09b84c215e4968d947a0a18f Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.811206 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5c0a-account-create-update-dmf7v" event={"ID":"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c","Type":"ContainerStarted","Data":"de40acca1612f9b201cd6ea209437a16f7894d4e3568935a93fe9074e1fff68b"} Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.815225 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-6kbkf" event={"ID":"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5","Type":"ContainerStarted","Data":"4a85a96da66480f8d3b1820febb85132ff07c41e09b84c215e4968d947a0a18f"} Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.818051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6654448665-7d6rf" event={"ID":"223e253a-8088-4569-935b-7f7bed328067","Type":"ContainerStarted","Data":"b461ccb9f17e51cf2946ef6f1b334b7c39e297972f166c7decf091ccdee3fc52"} Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.818080 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6654448665-7d6rf" event={"ID":"223e253a-8088-4569-935b-7f7bed328067","Type":"ContainerStarted","Data":"181d3b90268a52c434f5184234c155316670a5e75ed46cc7b6b472986752b280"} Dec 03 18:42:32 crc kubenswrapper[4758]: I1203 18:42:32.840741 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6654448665-7d6rf" podStartSLOduration=2.840726155 podStartE2EDuration="2.840726155s" podCreationTimestamp="2025-12-03 18:42:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:42:32.837390325 +0000 UTC m=+6408.038767186" watchObservedRunningTime="2025-12-03 18:42:32.840726155 +0000 UTC m=+6408.042103016" Dec 03 18:42:33 crc kubenswrapper[4758]: I1203 18:42:33.831149 4758 generic.go:334] "Generic (PLEG): container finished" podID="b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" containerID="4497e840ec067ef7d5272d120eeb4b11167c1074923c2dc3726155ac792c5ef2" exitCode=0 Dec 03 18:42:33 crc kubenswrapper[4758]: I1203 18:42:33.831244 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5c0a-account-create-update-dmf7v" event={"ID":"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c","Type":"ContainerDied","Data":"4497e840ec067ef7d5272d120eeb4b11167c1074923c2dc3726155ac792c5ef2"} Dec 03 18:42:33 crc kubenswrapper[4758]: I1203 18:42:33.835144 4758 generic.go:334] "Generic (PLEG): container finished" podID="6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" containerID="381a57e3ee67f4f00bb58b989420683a12c1a27a5fd06cec05ba15186632a020" exitCode=0 Dec 03 18:42:33 crc kubenswrapper[4758]: I1203 18:42:33.835231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-6kbkf" event={"ID":"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5","Type":"ContainerDied","Data":"381a57e3ee67f4f00bb58b989420683a12c1a27a5fd06cec05ba15186632a020"} Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.324094 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.335408 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.471628 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-operator-scripts\") pod \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.471728 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fg5c\" (UniqueName: \"kubernetes.io/projected/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-kube-api-access-9fg5c\") pod \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\" (UID: \"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c\") " Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.472257 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-operator-scripts\") pod \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.472366 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvq2v\" (UniqueName: \"kubernetes.io/projected/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-kube-api-access-pvq2v\") pod \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\" (UID: \"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5\") " Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.472946 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" (UID: "6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.473197 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" (UID: "b3e4a382-3f17-482b-a2c2-cbdc11dbef9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.479987 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-kube-api-access-pvq2v" (OuterVolumeSpecName: "kube-api-access-pvq2v") pod "6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" (UID: "6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5"). InnerVolumeSpecName "kube-api-access-pvq2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.481031 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-kube-api-access-9fg5c" (OuterVolumeSpecName: "kube-api-access-9fg5c") pod "b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" (UID: "b3e4a382-3f17-482b-a2c2-cbdc11dbef9c"). InnerVolumeSpecName "kube-api-access-9fg5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.575279 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fg5c\" (UniqueName: \"kubernetes.io/projected/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-kube-api-access-9fg5c\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.575313 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.575322 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvq2v\" (UniqueName: \"kubernetes.io/projected/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5-kube-api-access-pvq2v\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.575331 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.852394 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5c0a-account-create-update-dmf7v" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.852480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5c0a-account-create-update-dmf7v" event={"ID":"b3e4a382-3f17-482b-a2c2-cbdc11dbef9c","Type":"ContainerDied","Data":"de40acca1612f9b201cd6ea209437a16f7894d4e3568935a93fe9074e1fff68b"} Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.852516 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de40acca1612f9b201cd6ea209437a16f7894d4e3568935a93fe9074e1fff68b" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.854237 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-6kbkf" event={"ID":"6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5","Type":"ContainerDied","Data":"4a85a96da66480f8d3b1820febb85132ff07c41e09b84c215e4968d947a0a18f"} Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.854265 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a85a96da66480f8d3b1820febb85132ff07c41e09b84c215e4968d947a0a18f" Dec 03 18:42:35 crc kubenswrapper[4758]: I1203 18:42:35.854302 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-6kbkf" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.113475 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-k7jtl"] Dec 03 18:42:37 crc kubenswrapper[4758]: E1203 18:42:37.114385 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" containerName="mariadb-account-create-update" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.114402 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" containerName="mariadb-account-create-update" Dec 03 18:42:37 crc kubenswrapper[4758]: E1203 18:42:37.114445 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" containerName="mariadb-database-create" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.114457 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" containerName="mariadb-database-create" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.114770 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" containerName="mariadb-database-create" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.114795 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" containerName="mariadb-account-create-update" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.116018 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.117876 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-9mfx8" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.118999 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.154000 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-k7jtl"] Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.209981 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-combined-ca-bundle\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.210296 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-config-data\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.210397 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5thcq\" (UniqueName: \"kubernetes.io/projected/2205b3e1-3e92-4688-a666-a730dd1cce69-kube-api-access-5thcq\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.313114 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-config-data\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.313229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5thcq\" (UniqueName: \"kubernetes.io/projected/2205b3e1-3e92-4688-a666-a730dd1cce69-kube-api-access-5thcq\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.313322 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-combined-ca-bundle\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.321555 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-config-data\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.330234 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-combined-ca-bundle\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.337570 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5thcq\" (UniqueName: \"kubernetes.io/projected/2205b3e1-3e92-4688-a666-a730dd1cce69-kube-api-access-5thcq\") pod \"heat-db-sync-k7jtl\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.445496 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:37 crc kubenswrapper[4758]: I1203 18:42:37.913609 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-k7jtl"] Dec 03 18:42:38 crc kubenswrapper[4758]: I1203 18:42:38.887781 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-k7jtl" event={"ID":"2205b3e1-3e92-4688-a666-a730dd1cce69","Type":"ContainerStarted","Data":"b7c8b94b35a06d7f3beae45c4959dbf1ca48b3a2858502f609bf54ae535a2b74"} Dec 03 18:42:40 crc kubenswrapper[4758]: I1203 18:42:40.840254 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:40 crc kubenswrapper[4758]: I1203 18:42:40.842480 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:45 crc kubenswrapper[4758]: I1203 18:42:45.954962 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-k7jtl" event={"ID":"2205b3e1-3e92-4688-a666-a730dd1cce69","Type":"ContainerStarted","Data":"b39142ea20bffd6d24ece9d10ad0e515a58f35e65400ed8aad77a8ab4c287405"} Dec 03 18:42:45 crc kubenswrapper[4758]: I1203 18:42:45.976121 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-k7jtl" podStartSLOduration=1.616648193 podStartE2EDuration="8.976101787s" podCreationTimestamp="2025-12-03 18:42:37 +0000 UTC" firstStartedPulling="2025-12-03 18:42:37.924335023 +0000 UTC m=+6413.125711884" lastFinishedPulling="2025-12-03 18:42:45.283788607 +0000 UTC m=+6420.485165478" observedRunningTime="2025-12-03 18:42:45.970238957 +0000 UTC m=+6421.171615838" watchObservedRunningTime="2025-12-03 18:42:45.976101787 +0000 UTC m=+6421.177478648" Dec 03 18:42:47 crc kubenswrapper[4758]: I1203 18:42:47.977332 4758 generic.go:334] "Generic (PLEG): container finished" podID="2205b3e1-3e92-4688-a666-a730dd1cce69" containerID="b39142ea20bffd6d24ece9d10ad0e515a58f35e65400ed8aad77a8ab4c287405" exitCode=0 Dec 03 18:42:47 crc kubenswrapper[4758]: I1203 18:42:47.977378 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-k7jtl" event={"ID":"2205b3e1-3e92-4688-a666-a730dd1cce69","Type":"ContainerDied","Data":"b39142ea20bffd6d24ece9d10ad0e515a58f35e65400ed8aad77a8ab4c287405"} Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.394891 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.466031 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5thcq\" (UniqueName: \"kubernetes.io/projected/2205b3e1-3e92-4688-a666-a730dd1cce69-kube-api-access-5thcq\") pod \"2205b3e1-3e92-4688-a666-a730dd1cce69\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.466246 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-combined-ca-bundle\") pod \"2205b3e1-3e92-4688-a666-a730dd1cce69\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.466363 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-config-data\") pod \"2205b3e1-3e92-4688-a666-a730dd1cce69\" (UID: \"2205b3e1-3e92-4688-a666-a730dd1cce69\") " Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.472818 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2205b3e1-3e92-4688-a666-a730dd1cce69-kube-api-access-5thcq" (OuterVolumeSpecName: "kube-api-access-5thcq") pod "2205b3e1-3e92-4688-a666-a730dd1cce69" (UID: "2205b3e1-3e92-4688-a666-a730dd1cce69"). InnerVolumeSpecName "kube-api-access-5thcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.493418 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2205b3e1-3e92-4688-a666-a730dd1cce69" (UID: "2205b3e1-3e92-4688-a666-a730dd1cce69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.545449 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-config-data" (OuterVolumeSpecName: "config-data") pod "2205b3e1-3e92-4688-a666-a730dd1cce69" (UID: "2205b3e1-3e92-4688-a666-a730dd1cce69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.568065 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.568114 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2205b3e1-3e92-4688-a666-a730dd1cce69-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:49 crc kubenswrapper[4758]: I1203 18:42:49.568128 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5thcq\" (UniqueName: \"kubernetes.io/projected/2205b3e1-3e92-4688-a666-a730dd1cce69-kube-api-access-5thcq\") on node \"crc\" DevicePath \"\"" Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.004004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-k7jtl" event={"ID":"2205b3e1-3e92-4688-a666-a730dd1cce69","Type":"ContainerDied","Data":"b7c8b94b35a06d7f3beae45c4959dbf1ca48b3a2858502f609bf54ae535a2b74"} Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.004081 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7c8b94b35a06d7f3beae45c4959dbf1ca48b3a2858502f609bf54ae535a2b74" Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.004190 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-k7jtl" Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.994248 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6f84647496-588vd"] Dec 03 18:42:50 crc kubenswrapper[4758]: E1203 18:42:50.995025 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2205b3e1-3e92-4688-a666-a730dd1cce69" containerName="heat-db-sync" Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.995040 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2205b3e1-3e92-4688-a666-a730dd1cce69" containerName="heat-db-sync" Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.995266 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2205b3e1-3e92-4688-a666-a730dd1cce69" containerName="heat-db-sync" Dec 03 18:42:50 crc kubenswrapper[4758]: I1203 18:42:50.996251 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.007931 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-9mfx8" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.008594 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.009091 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.035379 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6f84647496-588vd"] Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.103220 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-config-data-custom\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.103267 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzwhc\" (UniqueName: \"kubernetes.io/projected/9121c214-9012-49c7-8f48-34f2327d3cb9-kube-api-access-lzwhc\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.103286 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-combined-ca-bundle\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.103304 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-config-data\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.176426 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-cd9dd694-vjtvk"] Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.177664 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.185940 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.205745 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-config-data-custom\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.205819 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzwhc\" (UniqueName: \"kubernetes.io/projected/9121c214-9012-49c7-8f48-34f2327d3cb9-kube-api-access-lzwhc\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.205851 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-combined-ca-bundle\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.205875 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-config-data\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.211331 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7bf88cbcf7-7tk6c"] Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.213047 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.215481 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.222393 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7bf88cbcf7-7tk6c"] Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.223952 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-combined-ca-bundle\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.224039 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-config-data-custom\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.224518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9121c214-9012-49c7-8f48-34f2327d3cb9-config-data\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.228429 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzwhc\" (UniqueName: \"kubernetes.io/projected/9121c214-9012-49c7-8f48-34f2327d3cb9-kube-api-access-lzwhc\") pod \"heat-engine-6f84647496-588vd\" (UID: \"9121c214-9012-49c7-8f48-34f2327d3cb9\") " pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.234758 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-cd9dd694-vjtvk"] Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.308236 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9jk2\" (UniqueName: \"kubernetes.io/projected/bede3802-5d16-4df3-8595-4af3c5be0272-kube-api-access-n9jk2\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.308640 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-combined-ca-bundle\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.308773 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-config-data-custom\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.308865 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-config-data\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.308906 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczbc\" (UniqueName: \"kubernetes.io/projected/508aa60b-f57e-46f2-97f4-43e7c7bedf31-kube-api-access-hczbc\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.308938 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-config-data\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.309040 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-combined-ca-bundle\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.309068 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-config-data-custom\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.329311 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412741 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-combined-ca-bundle\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412797 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-config-data-custom\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412828 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-config-data\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412849 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczbc\" (UniqueName: \"kubernetes.io/projected/508aa60b-f57e-46f2-97f4-43e7c7bedf31-kube-api-access-hczbc\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412869 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-config-data\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412907 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-combined-ca-bundle\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.412923 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-config-data-custom\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.413012 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9jk2\" (UniqueName: \"kubernetes.io/projected/bede3802-5d16-4df3-8595-4af3c5be0272-kube-api-access-n9jk2\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.419228 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-config-data-custom\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.422671 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-config-data\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.423439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-config-data-custom\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.425075 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-config-data\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.435527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczbc\" (UniqueName: \"kubernetes.io/projected/508aa60b-f57e-46f2-97f4-43e7c7bedf31-kube-api-access-hczbc\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.435938 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9jk2\" (UniqueName: \"kubernetes.io/projected/bede3802-5d16-4df3-8595-4af3c5be0272-kube-api-access-n9jk2\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.436432 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508aa60b-f57e-46f2-97f4-43e7c7bedf31-combined-ca-bundle\") pod \"heat-cfnapi-7bf88cbcf7-7tk6c\" (UID: \"508aa60b-f57e-46f2-97f4-43e7c7bedf31\") " pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.438504 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bede3802-5d16-4df3-8595-4af3c5be0272-combined-ca-bundle\") pod \"heat-api-cd9dd694-vjtvk\" (UID: \"bede3802-5d16-4df3-8595-4af3c5be0272\") " pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.500914 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.620388 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:51 crc kubenswrapper[4758]: I1203 18:42:51.854919 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6f84647496-588vd"] Dec 03 18:42:52 crc kubenswrapper[4758]: I1203 18:42:52.060495 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6f84647496-588vd" event={"ID":"9121c214-9012-49c7-8f48-34f2327d3cb9","Type":"ContainerStarted","Data":"aa00e540bc4e840c9c57b1224ef259aad85d9ccdaf4e043d5f66ac7de281be66"} Dec 03 18:42:52 crc kubenswrapper[4758]: I1203 18:42:52.066088 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-cd9dd694-vjtvk"] Dec 03 18:42:52 crc kubenswrapper[4758]: W1203 18:42:52.066804 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbede3802_5d16_4df3_8595_4af3c5be0272.slice/crio-407a51771e3dd2d7fd42fdceeae2d34f8eff72ce6b3977079a470be7747a3b14 WatchSource:0}: Error finding container 407a51771e3dd2d7fd42fdceeae2d34f8eff72ce6b3977079a470be7747a3b14: Status 404 returned error can't find the container with id 407a51771e3dd2d7fd42fdceeae2d34f8eff72ce6b3977079a470be7747a3b14 Dec 03 18:42:52 crc kubenswrapper[4758]: I1203 18:42:52.173989 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7bf88cbcf7-7tk6c"] Dec 03 18:42:53 crc kubenswrapper[4758]: I1203 18:42:53.084634 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" event={"ID":"508aa60b-f57e-46f2-97f4-43e7c7bedf31","Type":"ContainerStarted","Data":"9cbd0dc1aa2534289eb4165c792eb0059cbb367eba0db5bf8b50a4322212397e"} Dec 03 18:42:53 crc kubenswrapper[4758]: I1203 18:42:53.087033 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-cd9dd694-vjtvk" event={"ID":"bede3802-5d16-4df3-8595-4af3c5be0272","Type":"ContainerStarted","Data":"407a51771e3dd2d7fd42fdceeae2d34f8eff72ce6b3977079a470be7747a3b14"} Dec 03 18:42:53 crc kubenswrapper[4758]: I1203 18:42:53.089517 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6f84647496-588vd" event={"ID":"9121c214-9012-49c7-8f48-34f2327d3cb9","Type":"ContainerStarted","Data":"4a12b9868ff2cb4ffa51c4f87659aeebacc8d0ccc1fbdc74b72c51c9452114e8"} Dec 03 18:42:53 crc kubenswrapper[4758]: I1203 18:42:53.089643 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:42:53 crc kubenswrapper[4758]: I1203 18:42:53.169135 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:53 crc kubenswrapper[4758]: I1203 18:42:53.195703 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6f84647496-588vd" podStartSLOduration=3.195664213 podStartE2EDuration="3.195664213s" podCreationTimestamp="2025-12-03 18:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:42:53.108721178 +0000 UTC m=+6428.310098039" watchObservedRunningTime="2025-12-03 18:42:53.195664213 +0000 UTC m=+6428.397041084" Dec 03 18:42:55 crc kubenswrapper[4758]: I1203 18:42:55.052636 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6654448665-7d6rf" Dec 03 18:42:55 crc kubenswrapper[4758]: I1203 18:42:55.136924 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-df5696bb9-ckrnz"] Dec 03 18:42:55 crc kubenswrapper[4758]: I1203 18:42:55.137124 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-df5696bb9-ckrnz" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon-log" containerID="cri-o://ad22110c038980687e63440cd8ad0b1244374f2f6d96881a371a94c6993293b8" gracePeriod=30 Dec 03 18:42:55 crc kubenswrapper[4758]: I1203 18:42:55.137250 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-df5696bb9-ckrnz" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" containerID="cri-o://24776dc59fd5b45da37bd7cd1d0a8645987082b054e38c439bb4a3b272dd280e" gracePeriod=30 Dec 03 18:42:56 crc kubenswrapper[4758]: I1203 18:42:56.125061 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-cd9dd694-vjtvk" event={"ID":"bede3802-5d16-4df3-8595-4af3c5be0272","Type":"ContainerStarted","Data":"1b2bb4e6ea845ce0774ebe00925351ed62e57089be20aa7ee2564ba5006cf9ae"} Dec 03 18:42:56 crc kubenswrapper[4758]: I1203 18:42:56.127187 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:42:56 crc kubenswrapper[4758]: I1203 18:42:56.132343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" event={"ID":"508aa60b-f57e-46f2-97f4-43e7c7bedf31","Type":"ContainerStarted","Data":"da18776df77e35ca20b6f72c22d3cc5cc2d64eadebcbd65d9620b47a5209fcc5"} Dec 03 18:42:56 crc kubenswrapper[4758]: I1203 18:42:56.132861 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:42:56 crc kubenswrapper[4758]: I1203 18:42:56.142411 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-cd9dd694-vjtvk" podStartSLOduration=2.101565554 podStartE2EDuration="5.142394628s" podCreationTimestamp="2025-12-03 18:42:51 +0000 UTC" firstStartedPulling="2025-12-03 18:42:52.069351689 +0000 UTC m=+6427.270728550" lastFinishedPulling="2025-12-03 18:42:55.110180763 +0000 UTC m=+6430.311557624" observedRunningTime="2025-12-03 18:42:56.140178658 +0000 UTC m=+6431.341555519" watchObservedRunningTime="2025-12-03 18:42:56.142394628 +0000 UTC m=+6431.343771489" Dec 03 18:42:56 crc kubenswrapper[4758]: I1203 18:42:56.166569 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" podStartSLOduration=2.240481977 podStartE2EDuration="5.166552803s" podCreationTimestamp="2025-12-03 18:42:51 +0000 UTC" firstStartedPulling="2025-12-03 18:42:52.182206515 +0000 UTC m=+6427.383583376" lastFinishedPulling="2025-12-03 18:42:55.108277341 +0000 UTC m=+6430.309654202" observedRunningTime="2025-12-03 18:42:56.156774598 +0000 UTC m=+6431.358151459" watchObservedRunningTime="2025-12-03 18:42:56.166552803 +0000 UTC m=+6431.367929664" Dec 03 18:42:59 crc kubenswrapper[4758]: I1203 18:42:59.182599 4758 generic.go:334] "Generic (PLEG): container finished" podID="339a325b-d675-41c1-87cb-22b328569743" containerID="24776dc59fd5b45da37bd7cd1d0a8645987082b054e38c439bb4a3b272dd280e" exitCode=0 Dec 03 18:42:59 crc kubenswrapper[4758]: I1203 18:42:59.182694 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-df5696bb9-ckrnz" event={"ID":"339a325b-d675-41c1-87cb-22b328569743","Type":"ContainerDied","Data":"24776dc59fd5b45da37bd7cd1d0a8645987082b054e38c439bb4a3b272dd280e"} Dec 03 18:43:02 crc kubenswrapper[4758]: I1203 18:43:02.966388 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-cd9dd694-vjtvk" Dec 03 18:43:03 crc kubenswrapper[4758]: I1203 18:43:03.229136 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7bf88cbcf7-7tk6c" Dec 03 18:43:04 crc kubenswrapper[4758]: I1203 18:43:04.361202 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-df5696bb9-ckrnz" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 03 18:43:11 crc kubenswrapper[4758]: I1203 18:43:11.358295 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6f84647496-588vd" Dec 03 18:43:11 crc kubenswrapper[4758]: I1203 18:43:11.394714 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:43:11 crc kubenswrapper[4758]: I1203 18:43:11.394783 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:43:14 crc kubenswrapper[4758]: I1203 18:43:14.253050 4758 scope.go:117] "RemoveContainer" containerID="93a6fb1412c9cf87fcc7b623b4ab03be77937a1653811378ce7fdad5ab6c01f4" Dec 03 18:43:14 crc kubenswrapper[4758]: I1203 18:43:14.361057 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-df5696bb9-ckrnz" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 03 18:43:20 crc kubenswrapper[4758]: I1203 18:43:20.042403 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-9mv8v"] Dec 03 18:43:20 crc kubenswrapper[4758]: I1203 18:43:20.053247 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-9mv8v"] Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.027264 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-bea5-account-create-update-d5xv2"] Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.036180 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-knwsq"] Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.046982 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-bea5-account-create-update-d5xv2"] Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.055890 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-knwsq"] Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.126394 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e84029c-4acb-4f87-baee-86fff838a10d" path="/var/lib/kubelet/pods/4e84029c-4acb-4f87-baee-86fff838a10d/volumes" Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.127485 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ff9e71-648d-48e2-b428-ee5676abc858" path="/var/lib/kubelet/pods/c2ff9e71-648d-48e2-b428-ee5676abc858/volumes" Dec 03 18:43:21 crc kubenswrapper[4758]: I1203 18:43:21.128677 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f27d6cf1-c67d-4b5d-9a54-d1681c949841" path="/var/lib/kubelet/pods/f27d6cf1-c67d-4b5d-9a54-d1681c949841/volumes" Dec 03 18:43:22 crc kubenswrapper[4758]: I1203 18:43:22.032434 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cdcf-account-create-update-bftp6"] Dec 03 18:43:22 crc kubenswrapper[4758]: I1203 18:43:22.045893 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4cde-account-create-update-9bwcq"] Dec 03 18:43:22 crc kubenswrapper[4758]: I1203 18:43:22.059347 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-kct8l"] Dec 03 18:43:22 crc kubenswrapper[4758]: I1203 18:43:22.070030 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4cde-account-create-update-9bwcq"] Dec 03 18:43:22 crc kubenswrapper[4758]: I1203 18:43:22.077691 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cdcf-account-create-update-bftp6"] Dec 03 18:43:22 crc kubenswrapper[4758]: I1203 18:43:22.085305 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-kct8l"] Dec 03 18:43:23 crc kubenswrapper[4758]: I1203 18:43:23.127312 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353f3f28-3228-4a1b-9485-8b5f38247264" path="/var/lib/kubelet/pods/353f3f28-3228-4a1b-9485-8b5f38247264/volumes" Dec 03 18:43:23 crc kubenswrapper[4758]: I1203 18:43:23.129447 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3597b615-57a3-47b9-a718-35e266252642" path="/var/lib/kubelet/pods/3597b615-57a3-47b9-a718-35e266252642/volumes" Dec 03 18:43:23 crc kubenswrapper[4758]: I1203 18:43:23.131786 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c270177-5d19-4a1a-a965-f2b106c43e26" path="/var/lib/kubelet/pods/5c270177-5d19-4a1a-a965-f2b106c43e26/volumes" Dec 03 18:43:24 crc kubenswrapper[4758]: I1203 18:43:24.361722 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-df5696bb9-ckrnz" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 03 18:43:24 crc kubenswrapper[4758]: I1203 18:43:24.362208 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.447462 4758 generic.go:334] "Generic (PLEG): container finished" podID="339a325b-d675-41c1-87cb-22b328569743" containerID="ad22110c038980687e63440cd8ad0b1244374f2f6d96881a371a94c6993293b8" exitCode=137 Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.447781 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-df5696bb9-ckrnz" event={"ID":"339a325b-d675-41c1-87cb-22b328569743","Type":"ContainerDied","Data":"ad22110c038980687e63440cd8ad0b1244374f2f6d96881a371a94c6993293b8"} Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.582224 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.704022 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4vw6\" (UniqueName: \"kubernetes.io/projected/339a325b-d675-41c1-87cb-22b328569743-kube-api-access-f4vw6\") pod \"339a325b-d675-41c1-87cb-22b328569743\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.704215 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-scripts\") pod \"339a325b-d675-41c1-87cb-22b328569743\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.704346 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-config-data\") pod \"339a325b-d675-41c1-87cb-22b328569743\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.704401 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339a325b-d675-41c1-87cb-22b328569743-logs\") pod \"339a325b-d675-41c1-87cb-22b328569743\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.704452 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/339a325b-d675-41c1-87cb-22b328569743-horizon-secret-key\") pod \"339a325b-d675-41c1-87cb-22b328569743\" (UID: \"339a325b-d675-41c1-87cb-22b328569743\") " Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.704831 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/339a325b-d675-41c1-87cb-22b328569743-logs" (OuterVolumeSpecName: "logs") pod "339a325b-d675-41c1-87cb-22b328569743" (UID: "339a325b-d675-41c1-87cb-22b328569743"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.705112 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339a325b-d675-41c1-87cb-22b328569743-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.733145 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339a325b-d675-41c1-87cb-22b328569743-kube-api-access-f4vw6" (OuterVolumeSpecName: "kube-api-access-f4vw6") pod "339a325b-d675-41c1-87cb-22b328569743" (UID: "339a325b-d675-41c1-87cb-22b328569743"). InnerVolumeSpecName "kube-api-access-f4vw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.749525 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339a325b-d675-41c1-87cb-22b328569743-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "339a325b-d675-41c1-87cb-22b328569743" (UID: "339a325b-d675-41c1-87cb-22b328569743"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.786571 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-scripts" (OuterVolumeSpecName: "scripts") pod "339a325b-d675-41c1-87cb-22b328569743" (UID: "339a325b-d675-41c1-87cb-22b328569743"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.791355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-config-data" (OuterVolumeSpecName: "config-data") pod "339a325b-d675-41c1-87cb-22b328569743" (UID: "339a325b-d675-41c1-87cb-22b328569743"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.807066 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.807108 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/339a325b-d675-41c1-87cb-22b328569743-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.807123 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4vw6\" (UniqueName: \"kubernetes.io/projected/339a325b-d675-41c1-87cb-22b328569743-kube-api-access-f4vw6\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:25 crc kubenswrapper[4758]: I1203 18:43:25.807135 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/339a325b-d675-41c1-87cb-22b328569743-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:26 crc kubenswrapper[4758]: I1203 18:43:26.461148 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-df5696bb9-ckrnz" event={"ID":"339a325b-d675-41c1-87cb-22b328569743","Type":"ContainerDied","Data":"42c82137443eaad9b418e57b73c2d6ef6e45d60309dd37079e872e995b2a11c8"} Dec 03 18:43:26 crc kubenswrapper[4758]: I1203 18:43:26.461206 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-df5696bb9-ckrnz" Dec 03 18:43:26 crc kubenswrapper[4758]: I1203 18:43:26.461246 4758 scope.go:117] "RemoveContainer" containerID="24776dc59fd5b45da37bd7cd1d0a8645987082b054e38c439bb4a3b272dd280e" Dec 03 18:43:26 crc kubenswrapper[4758]: I1203 18:43:26.508506 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-df5696bb9-ckrnz"] Dec 03 18:43:26 crc kubenswrapper[4758]: I1203 18:43:26.520091 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-df5696bb9-ckrnz"] Dec 03 18:43:26 crc kubenswrapper[4758]: I1203 18:43:26.644234 4758 scope.go:117] "RemoveContainer" containerID="ad22110c038980687e63440cd8ad0b1244374f2f6d96881a371a94c6993293b8" Dec 03 18:43:27 crc kubenswrapper[4758]: I1203 18:43:27.128275 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339a325b-d675-41c1-87cb-22b328569743" path="/var/lib/kubelet/pods/339a325b-d675-41c1-87cb-22b328569743/volumes" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.131347 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5"] Dec 03 18:43:30 crc kubenswrapper[4758]: E1203 18:43:30.132323 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon-log" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.132347 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon-log" Dec 03 18:43:30 crc kubenswrapper[4758]: E1203 18:43:30.132411 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.132423 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.132845 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.132884 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="339a325b-d675-41c1-87cb-22b328569743" containerName="horizon-log" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.135472 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.138205 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.140879 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5"] Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.298069 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f476x\" (UniqueName: \"kubernetes.io/projected/94b4bf6d-75a9-4b05-bb33-039174ec7a30-kube-api-access-f476x\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.298391 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.298614 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.400712 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.400861 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.400967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f476x\" (UniqueName: \"kubernetes.io/projected/94b4bf6d-75a9-4b05-bb33-039174ec7a30-kube-api-access-f476x\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.401669 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.401812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.426984 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f476x\" (UniqueName: \"kubernetes.io/projected/94b4bf6d-75a9-4b05-bb33-039174ec7a30-kube-api-access-f476x\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.465911 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:30 crc kubenswrapper[4758]: I1203 18:43:30.942933 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5"] Dec 03 18:43:31 crc kubenswrapper[4758]: I1203 18:43:31.517075 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" event={"ID":"94b4bf6d-75a9-4b05-bb33-039174ec7a30","Type":"ContainerStarted","Data":"b07fbdc89fd21ef5d4bb8963438142af752a1bf9aeee7f7037daa28dcfd8fe9e"} Dec 03 18:43:31 crc kubenswrapper[4758]: I1203 18:43:31.517146 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" event={"ID":"94b4bf6d-75a9-4b05-bb33-039174ec7a30","Type":"ContainerStarted","Data":"247c3fd9c1f5d6397c8e5839056d54df3f726dd664c38358fb6f6895b8b50ab4"} Dec 03 18:43:32 crc kubenswrapper[4758]: I1203 18:43:32.528421 4758 generic.go:334] "Generic (PLEG): container finished" podID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerID="b07fbdc89fd21ef5d4bb8963438142af752a1bf9aeee7f7037daa28dcfd8fe9e" exitCode=0 Dec 03 18:43:32 crc kubenswrapper[4758]: I1203 18:43:32.528479 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" event={"ID":"94b4bf6d-75a9-4b05-bb33-039174ec7a30","Type":"ContainerDied","Data":"b07fbdc89fd21ef5d4bb8963438142af752a1bf9aeee7f7037daa28dcfd8fe9e"} Dec 03 18:43:32 crc kubenswrapper[4758]: I1203 18:43:32.531076 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.453999 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xsmwc"] Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.456699 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.466656 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xsmwc"] Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.567961 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-utilities\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.568051 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h29hz\" (UniqueName: \"kubernetes.io/projected/ba406b4a-4146-4bf5-ab95-71a881d14143-kube-api-access-h29hz\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.568074 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-catalog-content\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.669822 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-utilities\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.669876 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h29hz\" (UniqueName: \"kubernetes.io/projected/ba406b4a-4146-4bf5-ab95-71a881d14143-kube-api-access-h29hz\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.669900 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-catalog-content\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.670414 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-catalog-content\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.670701 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-utilities\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.789113 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h29hz\" (UniqueName: \"kubernetes.io/projected/ba406b4a-4146-4bf5-ab95-71a881d14143-kube-api-access-h29hz\") pod \"certified-operators-xsmwc\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:33 crc kubenswrapper[4758]: I1203 18:43:33.790288 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:34 crc kubenswrapper[4758]: I1203 18:43:34.420629 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xsmwc"] Dec 03 18:43:34 crc kubenswrapper[4758]: W1203 18:43:34.431579 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba406b4a_4146_4bf5_ab95_71a881d14143.slice/crio-59324328bc92d9f80c4a1317086af83559e6a8d86bea18e5ab9fd8b4477b0de5 WatchSource:0}: Error finding container 59324328bc92d9f80c4a1317086af83559e6a8d86bea18e5ab9fd8b4477b0de5: Status 404 returned error can't find the container with id 59324328bc92d9f80c4a1317086af83559e6a8d86bea18e5ab9fd8b4477b0de5 Dec 03 18:43:34 crc kubenswrapper[4758]: I1203 18:43:34.559712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsmwc" event={"ID":"ba406b4a-4146-4bf5-ab95-71a881d14143","Type":"ContainerStarted","Data":"59324328bc92d9f80c4a1317086af83559e6a8d86bea18e5ab9fd8b4477b0de5"} Dec 03 18:43:35 crc kubenswrapper[4758]: I1203 18:43:35.571026 4758 generic.go:334] "Generic (PLEG): container finished" podID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerID="853fd350be11914b3989920bb041578518ffaaa20d7c127da76cfcb7257669bf" exitCode=0 Dec 03 18:43:35 crc kubenswrapper[4758]: I1203 18:43:35.571100 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" event={"ID":"94b4bf6d-75a9-4b05-bb33-039174ec7a30","Type":"ContainerDied","Data":"853fd350be11914b3989920bb041578518ffaaa20d7c127da76cfcb7257669bf"} Dec 03 18:43:35 crc kubenswrapper[4758]: I1203 18:43:35.573583 4758 generic.go:334] "Generic (PLEG): container finished" podID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerID="135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5" exitCode=0 Dec 03 18:43:35 crc kubenswrapper[4758]: I1203 18:43:35.573638 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsmwc" event={"ID":"ba406b4a-4146-4bf5-ab95-71a881d14143","Type":"ContainerDied","Data":"135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5"} Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.057560 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wvw82"] Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.072313 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wvw82"] Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.588755 4758 generic.go:334] "Generic (PLEG): container finished" podID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerID="b756e7f6a50093510bba006221b89ac0a2b383ab927ec2ef1e1033dfc2eff356" exitCode=0 Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.589023 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" event={"ID":"94b4bf6d-75a9-4b05-bb33-039174ec7a30","Type":"ContainerDied","Data":"b756e7f6a50093510bba006221b89ac0a2b383ab927ec2ef1e1033dfc2eff356"} Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.650132 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl67"] Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.652578 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.660934 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl67"] Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.744830 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-catalog-content\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.745013 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rnj\" (UniqueName: \"kubernetes.io/projected/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-kube-api-access-85rnj\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.745069 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-utilities\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.846657 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rnj\" (UniqueName: \"kubernetes.io/projected/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-kube-api-access-85rnj\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.846757 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-utilities\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.846798 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-catalog-content\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.847318 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-catalog-content\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:36 crc kubenswrapper[4758]: I1203 18:43:36.847540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-utilities\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.011986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rnj\" (UniqueName: \"kubernetes.io/projected/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-kube-api-access-85rnj\") pod \"redhat-marketplace-5hl67\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.129604 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f5e5bc-c00d-4cbf-90af-0d58b19eb613" path="/var/lib/kubelet/pods/b9f5e5bc-c00d-4cbf-90af-0d58b19eb613/volumes" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.279612 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.599040 4758 generic.go:334] "Generic (PLEG): container finished" podID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerID="8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f" exitCode=0 Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.599090 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsmwc" event={"ID":"ba406b4a-4146-4bf5-ab95-71a881d14143","Type":"ContainerDied","Data":"8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f"} Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.723328 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl67"] Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.905155 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.966862 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-util\") pod \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.966921 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f476x\" (UniqueName: \"kubernetes.io/projected/94b4bf6d-75a9-4b05-bb33-039174ec7a30-kube-api-access-f476x\") pod \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.967011 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-bundle\") pod \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\" (UID: \"94b4bf6d-75a9-4b05-bb33-039174ec7a30\") " Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.973130 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-bundle" (OuterVolumeSpecName: "bundle") pod "94b4bf6d-75a9-4b05-bb33-039174ec7a30" (UID: "94b4bf6d-75a9-4b05-bb33-039174ec7a30"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.976002 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b4bf6d-75a9-4b05-bb33-039174ec7a30-kube-api-access-f476x" (OuterVolumeSpecName: "kube-api-access-f476x") pod "94b4bf6d-75a9-4b05-bb33-039174ec7a30" (UID: "94b4bf6d-75a9-4b05-bb33-039174ec7a30"). InnerVolumeSpecName "kube-api-access-f476x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:43:37 crc kubenswrapper[4758]: I1203 18:43:37.976409 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-util" (OuterVolumeSpecName: "util") pod "94b4bf6d-75a9-4b05-bb33-039174ec7a30" (UID: "94b4bf6d-75a9-4b05-bb33-039174ec7a30"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.070400 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-util\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.070429 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f476x\" (UniqueName: \"kubernetes.io/projected/94b4bf6d-75a9-4b05-bb33-039174ec7a30-kube-api-access-f476x\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.070442 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94b4bf6d-75a9-4b05-bb33-039174ec7a30-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.614195 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" event={"ID":"94b4bf6d-75a9-4b05-bb33-039174ec7a30","Type":"ContainerDied","Data":"247c3fd9c1f5d6397c8e5839056d54df3f726dd664c38358fb6f6895b8b50ab4"} Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.614586 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="247c3fd9c1f5d6397c8e5839056d54df3f726dd664c38358fb6f6895b8b50ab4" Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.614224 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5" Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.617252 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsmwc" event={"ID":"ba406b4a-4146-4bf5-ab95-71a881d14143","Type":"ContainerStarted","Data":"e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591"} Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.620242 4758 generic.go:334] "Generic (PLEG): container finished" podID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerID="cd75fc4c6dee63f24256f4ddac9370009ab5dcc7690b00c04308cae27e85f227" exitCode=0 Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.620286 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl67" event={"ID":"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2","Type":"ContainerDied","Data":"cd75fc4c6dee63f24256f4ddac9370009ab5dcc7690b00c04308cae27e85f227"} Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.620310 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl67" event={"ID":"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2","Type":"ContainerStarted","Data":"2703e5b398bfb21f10bb13b2d58facdcf2d9b95ebc8d72f2705273c860c36695"} Dec 03 18:43:38 crc kubenswrapper[4758]: I1203 18:43:38.653624 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xsmwc" podStartSLOduration=3.15892662 podStartE2EDuration="5.653606851s" podCreationTimestamp="2025-12-03 18:43:33 +0000 UTC" firstStartedPulling="2025-12-03 18:43:35.576485116 +0000 UTC m=+6470.777861977" lastFinishedPulling="2025-12-03 18:43:38.071165347 +0000 UTC m=+6473.272542208" observedRunningTime="2025-12-03 18:43:38.636803468 +0000 UTC m=+6473.838180329" watchObservedRunningTime="2025-12-03 18:43:38.653606851 +0000 UTC m=+6473.854983712" Dec 03 18:43:40 crc kubenswrapper[4758]: I1203 18:43:40.647627 4758 generic.go:334] "Generic (PLEG): container finished" podID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerID="5c340af146e477c77044436457a21f57c92f51f43d0f32097dd7cedaddb750b7" exitCode=0 Dec 03 18:43:40 crc kubenswrapper[4758]: I1203 18:43:40.647801 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl67" event={"ID":"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2","Type":"ContainerDied","Data":"5c340af146e477c77044436457a21f57c92f51f43d0f32097dd7cedaddb750b7"} Dec 03 18:43:41 crc kubenswrapper[4758]: I1203 18:43:41.394668 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:43:41 crc kubenswrapper[4758]: I1203 18:43:41.395168 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:43:41 crc kubenswrapper[4758]: I1203 18:43:41.675290 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl67" event={"ID":"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2","Type":"ContainerStarted","Data":"0be6a4a0f5f48c33ab9173badcabd6c2822de5d3093ff7bf485abf420ddfb4fa"} Dec 03 18:43:41 crc kubenswrapper[4758]: I1203 18:43:41.698406 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5hl67" podStartSLOduration=3.223896028 podStartE2EDuration="5.698388196s" podCreationTimestamp="2025-12-03 18:43:36 +0000 UTC" firstStartedPulling="2025-12-03 18:43:38.623625243 +0000 UTC m=+6473.825002104" lastFinishedPulling="2025-12-03 18:43:41.098117411 +0000 UTC m=+6476.299494272" observedRunningTime="2025-12-03 18:43:41.697535532 +0000 UTC m=+6476.898912393" watchObservedRunningTime="2025-12-03 18:43:41.698388196 +0000 UTC m=+6476.899765057" Dec 03 18:43:43 crc kubenswrapper[4758]: I1203 18:43:43.790486 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:43 crc kubenswrapper[4758]: I1203 18:43:43.791786 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:43 crc kubenswrapper[4758]: I1203 18:43:43.847311 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:44 crc kubenswrapper[4758]: I1203 18:43:44.805225 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:46 crc kubenswrapper[4758]: I1203 18:43:46.441570 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xsmwc"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.281953 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.282002 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.336276 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.496369 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8"] Dec 03 18:43:47 crc kubenswrapper[4758]: E1203 18:43:47.560050 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="extract" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.560088 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="extract" Dec 03 18:43:47 crc kubenswrapper[4758]: E1203 18:43:47.560129 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="pull" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.560138 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="pull" Dec 03 18:43:47 crc kubenswrapper[4758]: E1203 18:43:47.560153 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="util" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.560160 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="util" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.560566 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b4bf6d-75a9-4b05-bb33-039174ec7a30" containerName="extract" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.561340 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.561428 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.565805 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.566005 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.566637 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-hpnkt" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.580719 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.582104 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.584135 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-k5zld" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.584698 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.593792 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.593881 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2mrq\" (UniqueName: \"kubernetes.io/projected/373f7373-1c0a-4261-81b5-03a8efcf147e-kube-api-access-x2mrq\") pod \"obo-prometheus-operator-668cf9dfbb-rkkv8\" (UID: \"373f7373-1c0a-4261-81b5-03a8efcf147e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.602629 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.605376 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.620927 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.697987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aa3f69cb-5012-4879-864e-5826e71f2aa8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-tgltn\" (UID: \"aa3f69cb-5012-4879-864e-5826e71f2aa8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.698112 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/18a96e71-fd5d-4816-a001-92ae35d94c17-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b\" (UID: \"18a96e71-fd5d-4816-a001-92ae35d94c17\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.698259 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aa3f69cb-5012-4879-864e-5826e71f2aa8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-tgltn\" (UID: \"aa3f69cb-5012-4879-864e-5826e71f2aa8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.698459 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2mrq\" (UniqueName: \"kubernetes.io/projected/373f7373-1c0a-4261-81b5-03a8efcf147e-kube-api-access-x2mrq\") pod \"obo-prometheus-operator-668cf9dfbb-rkkv8\" (UID: \"373f7373-1c0a-4261-81b5-03a8efcf147e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.698512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/18a96e71-fd5d-4816-a001-92ae35d94c17-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b\" (UID: \"18a96e71-fd5d-4816-a001-92ae35d94c17\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.713712 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-r27q7"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.715179 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.730209 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-vrvh4" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.731811 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.739769 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-r27q7"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.750896 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2mrq\" (UniqueName: \"kubernetes.io/projected/373f7373-1c0a-4261-81b5-03a8efcf147e-kube-api-access-x2mrq\") pod \"obo-prometheus-operator-668cf9dfbb-rkkv8\" (UID: \"373f7373-1c0a-4261-81b5-03a8efcf147e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.779887 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xsmwc" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="registry-server" containerID="cri-o://e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591" gracePeriod=2 Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.799743 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aa3f69cb-5012-4879-864e-5826e71f2aa8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-tgltn\" (UID: \"aa3f69cb-5012-4879-864e-5826e71f2aa8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.800067 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/18a96e71-fd5d-4816-a001-92ae35d94c17-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b\" (UID: \"18a96e71-fd5d-4816-a001-92ae35d94c17\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.800120 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aa3f69cb-5012-4879-864e-5826e71f2aa8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-tgltn\" (UID: \"aa3f69cb-5012-4879-864e-5826e71f2aa8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.800211 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd43e0c9-4231-459e-8d32-eddb92573c5b-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-r27q7\" (UID: \"dd43e0c9-4231-459e-8d32-eddb92573c5b\") " pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.800242 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlrf2\" (UniqueName: \"kubernetes.io/projected/dd43e0c9-4231-459e-8d32-eddb92573c5b-kube-api-access-jlrf2\") pod \"observability-operator-d8bb48f5d-r27q7\" (UID: \"dd43e0c9-4231-459e-8d32-eddb92573c5b\") " pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.800271 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/18a96e71-fd5d-4816-a001-92ae35d94c17-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b\" (UID: \"18a96e71-fd5d-4816-a001-92ae35d94c17\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.804987 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/18a96e71-fd5d-4816-a001-92ae35d94c17-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b\" (UID: \"18a96e71-fd5d-4816-a001-92ae35d94c17\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.805295 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aa3f69cb-5012-4879-864e-5826e71f2aa8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-tgltn\" (UID: \"aa3f69cb-5012-4879-864e-5826e71f2aa8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.806388 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/18a96e71-fd5d-4816-a001-92ae35d94c17-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b\" (UID: \"18a96e71-fd5d-4816-a001-92ae35d94c17\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.827409 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aa3f69cb-5012-4879-864e-5826e71f2aa8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6976796f4b-tgltn\" (UID: \"aa3f69cb-5012-4879-864e-5826e71f2aa8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.859762 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.892765 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.903009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd43e0c9-4231-459e-8d32-eddb92573c5b-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-r27q7\" (UID: \"dd43e0c9-4231-459e-8d32-eddb92573c5b\") " pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.903840 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlrf2\" (UniqueName: \"kubernetes.io/projected/dd43e0c9-4231-459e-8d32-eddb92573c5b-kube-api-access-jlrf2\") pod \"observability-operator-d8bb48f5d-r27q7\" (UID: \"dd43e0c9-4231-459e-8d32-eddb92573c5b\") " pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.910986 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd43e0c9-4231-459e-8d32-eddb92573c5b-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-r27q7\" (UID: \"dd43e0c9-4231-459e-8d32-eddb92573c5b\") " pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.913061 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.930634 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-h79l8"] Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.932169 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.941352 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-cth6g" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.943138 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.957947 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlrf2\" (UniqueName: \"kubernetes.io/projected/dd43e0c9-4231-459e-8d32-eddb92573c5b-kube-api-access-jlrf2\") pod \"observability-operator-d8bb48f5d-r27q7\" (UID: \"dd43e0c9-4231-459e-8d32-eddb92573c5b\") " pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:47 crc kubenswrapper[4758]: I1203 18:43:47.991757 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-h79l8"] Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.006141 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/424f6469-230e-4f6b-b257-7ad7feecd589-openshift-service-ca\") pod \"perses-operator-5446b9c989-h79l8\" (UID: \"424f6469-230e-4f6b-b257-7ad7feecd589\") " pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.006323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqghm\" (UniqueName: \"kubernetes.io/projected/424f6469-230e-4f6b-b257-7ad7feecd589-kube-api-access-hqghm\") pod \"perses-operator-5446b9c989-h79l8\" (UID: \"424f6469-230e-4f6b-b257-7ad7feecd589\") " pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.098989 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.108666 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqghm\" (UniqueName: \"kubernetes.io/projected/424f6469-230e-4f6b-b257-7ad7feecd589-kube-api-access-hqghm\") pod \"perses-operator-5446b9c989-h79l8\" (UID: \"424f6469-230e-4f6b-b257-7ad7feecd589\") " pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.109147 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/424f6469-230e-4f6b-b257-7ad7feecd589-openshift-service-ca\") pod \"perses-operator-5446b9c989-h79l8\" (UID: \"424f6469-230e-4f6b-b257-7ad7feecd589\") " pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.110276 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/424f6469-230e-4f6b-b257-7ad7feecd589-openshift-service-ca\") pod \"perses-operator-5446b9c989-h79l8\" (UID: \"424f6469-230e-4f6b-b257-7ad7feecd589\") " pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.156110 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqghm\" (UniqueName: \"kubernetes.io/projected/424f6469-230e-4f6b-b257-7ad7feecd589-kube-api-access-hqghm\") pod \"perses-operator-5446b9c989-h79l8\" (UID: \"424f6469-230e-4f6b-b257-7ad7feecd589\") " pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.421258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.455795 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.521176 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h29hz\" (UniqueName: \"kubernetes.io/projected/ba406b4a-4146-4bf5-ab95-71a881d14143-kube-api-access-h29hz\") pod \"ba406b4a-4146-4bf5-ab95-71a881d14143\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.521251 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-utilities\") pod \"ba406b4a-4146-4bf5-ab95-71a881d14143\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.521276 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-catalog-content\") pod \"ba406b4a-4146-4bf5-ab95-71a881d14143\" (UID: \"ba406b4a-4146-4bf5-ab95-71a881d14143\") " Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.526308 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-utilities" (OuterVolumeSpecName: "utilities") pod "ba406b4a-4146-4bf5-ab95-71a881d14143" (UID: "ba406b4a-4146-4bf5-ab95-71a881d14143"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.531248 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba406b4a-4146-4bf5-ab95-71a881d14143-kube-api-access-h29hz" (OuterVolumeSpecName: "kube-api-access-h29hz") pod "ba406b4a-4146-4bf5-ab95-71a881d14143" (UID: "ba406b4a-4146-4bf5-ab95-71a881d14143"). InnerVolumeSpecName "kube-api-access-h29hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.604211 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba406b4a-4146-4bf5-ab95-71a881d14143" (UID: "ba406b4a-4146-4bf5-ab95-71a881d14143"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.623855 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.623890 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba406b4a-4146-4bf5-ab95-71a881d14143-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.623906 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h29hz\" (UniqueName: \"kubernetes.io/projected/ba406b4a-4146-4bf5-ab95-71a881d14143-kube-api-access-h29hz\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.843935 4758 generic.go:334] "Generic (PLEG): container finished" podID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerID="e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591" exitCode=0 Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.844142 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsmwc" Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.844118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsmwc" event={"ID":"ba406b4a-4146-4bf5-ab95-71a881d14143","Type":"ContainerDied","Data":"e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591"} Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.844216 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsmwc" event={"ID":"ba406b4a-4146-4bf5-ab95-71a881d14143","Type":"ContainerDied","Data":"59324328bc92d9f80c4a1317086af83559e6a8d86bea18e5ab9fd8b4477b0de5"} Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.844248 4758 scope.go:117] "RemoveContainer" containerID="e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591" Dec 03 18:43:48 crc kubenswrapper[4758]: W1203 18:43:48.869484 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod373f7373_1c0a_4261_81b5_03a8efcf147e.slice/crio-8bb0a2cd1efbec9e5f83b0918a91ad79681c686dd3e73c343978e8db8cc6a534 WatchSource:0}: Error finding container 8bb0a2cd1efbec9e5f83b0918a91ad79681c686dd3e73c343978e8db8cc6a534: Status 404 returned error can't find the container with id 8bb0a2cd1efbec9e5f83b0918a91ad79681c686dd3e73c343978e8db8cc6a534 Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.894725 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8"] Dec 03 18:43:48 crc kubenswrapper[4758]: I1203 18:43:48.969386 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b"] Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.038327 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xsmwc"] Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.038994 4758 scope.go:117] "RemoveContainer" containerID="8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.046155 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xsmwc"] Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.098031 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn"] Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.132576 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" path="/var/lib/kubelet/pods/ba406b4a-4146-4bf5-ab95-71a881d14143/volumes" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.134966 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-h79l8"] Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.153663 4758 scope.go:117] "RemoveContainer" containerID="135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.197314 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-r27q7"] Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.228144 4758 scope.go:117] "RemoveContainer" containerID="e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591" Dec 03 18:43:49 crc kubenswrapper[4758]: E1203 18:43:49.230779 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591\": container with ID starting with e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591 not found: ID does not exist" containerID="e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.230832 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591"} err="failed to get container status \"e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591\": rpc error: code = NotFound desc = could not find container \"e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591\": container with ID starting with e21e45ebdbfe4ef3dabbf870e82034c7bbab3b665790974ea864f911a732f591 not found: ID does not exist" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.230861 4758 scope.go:117] "RemoveContainer" containerID="8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f" Dec 03 18:43:49 crc kubenswrapper[4758]: E1203 18:43:49.232869 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f\": container with ID starting with 8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f not found: ID does not exist" containerID="8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.232919 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f"} err="failed to get container status \"8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f\": rpc error: code = NotFound desc = could not find container \"8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f\": container with ID starting with 8d17a93f8616d769728e8797c8408203744c76af863219c647ef8f5d1160c81f not found: ID does not exist" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.232945 4758 scope.go:117] "RemoveContainer" containerID="135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5" Dec 03 18:43:49 crc kubenswrapper[4758]: E1203 18:43:49.233459 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5\": container with ID starting with 135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5 not found: ID does not exist" containerID="135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.233506 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5"} err="failed to get container status \"135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5\": rpc error: code = NotFound desc = could not find container \"135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5\": container with ID starting with 135793bfed14a76bdd55985eef9fa2a9b946238cefc6f1d2f71f6ae28a3c30c5 not found: ID does not exist" Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.854921 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" event={"ID":"373f7373-1c0a-4261-81b5-03a8efcf147e","Type":"ContainerStarted","Data":"8bb0a2cd1efbec9e5f83b0918a91ad79681c686dd3e73c343978e8db8cc6a534"} Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.856243 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-h79l8" event={"ID":"424f6469-230e-4f6b-b257-7ad7feecd589","Type":"ContainerStarted","Data":"3e9e6a0e5988f19227a626c6d9553156da83803428dd23c7f8aa6692a7949c61"} Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.857985 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" event={"ID":"dd43e0c9-4231-459e-8d32-eddb92573c5b","Type":"ContainerStarted","Data":"11478681c5afa6548bd96f2bdd2775c21baffd9ccd82534bd611f0aa227453a4"} Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.859339 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" event={"ID":"18a96e71-fd5d-4816-a001-92ae35d94c17","Type":"ContainerStarted","Data":"7625b2b044eae45b07257037fdf2c34372781a38dc80f805a0a643363bb4a5b8"} Dec 03 18:43:49 crc kubenswrapper[4758]: I1203 18:43:49.860303 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" event={"ID":"aa3f69cb-5012-4879-864e-5826e71f2aa8","Type":"ContainerStarted","Data":"fa7967904f881ab108f87a0ba7e81890f411b3e8d5ce56120d2782770ac7a2ca"} Dec 03 18:43:50 crc kubenswrapper[4758]: I1203 18:43:50.032757 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h6pbk"] Dec 03 18:43:50 crc kubenswrapper[4758]: I1203 18:43:50.043494 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h6pbk"] Dec 03 18:43:50 crc kubenswrapper[4758]: I1203 18:43:50.841138 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl67"] Dec 03 18:43:50 crc kubenswrapper[4758]: I1203 18:43:50.841698 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5hl67" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="registry-server" containerID="cri-o://0be6a4a0f5f48c33ab9173badcabd6c2822de5d3093ff7bf485abf420ddfb4fa" gracePeriod=2 Dec 03 18:43:51 crc kubenswrapper[4758]: I1203 18:43:51.052465 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlql4"] Dec 03 18:43:51 crc kubenswrapper[4758]: I1203 18:43:51.067650 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-wlql4"] Dec 03 18:43:51 crc kubenswrapper[4758]: I1203 18:43:51.128377 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5dd0a8-1259-4bca-b035-485c40718803" path="/var/lib/kubelet/pods/8a5dd0a8-1259-4bca-b035-485c40718803/volumes" Dec 03 18:43:51 crc kubenswrapper[4758]: I1203 18:43:51.129574 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a87223ea-1a03-43c6-bbe0-e9a11b4db504" path="/var/lib/kubelet/pods/a87223ea-1a03-43c6-bbe0-e9a11b4db504/volumes" Dec 03 18:43:51 crc kubenswrapper[4758]: I1203 18:43:51.914755 4758 generic.go:334] "Generic (PLEG): container finished" podID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerID="0be6a4a0f5f48c33ab9173badcabd6c2822de5d3093ff7bf485abf420ddfb4fa" exitCode=0 Dec 03 18:43:51 crc kubenswrapper[4758]: I1203 18:43:51.915092 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl67" event={"ID":"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2","Type":"ContainerDied","Data":"0be6a4a0f5f48c33ab9173badcabd6c2822de5d3093ff7bf485abf420ddfb4fa"} Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.353979 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.473007 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-utilities\") pod \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.473422 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-catalog-content\") pod \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.473478 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85rnj\" (UniqueName: \"kubernetes.io/projected/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-kube-api-access-85rnj\") pod \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\" (UID: \"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2\") " Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.473816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-utilities" (OuterVolumeSpecName: "utilities") pod "e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" (UID: "e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.474377 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.486484 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-kube-api-access-85rnj" (OuterVolumeSpecName: "kube-api-access-85rnj") pod "e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" (UID: "e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2"). InnerVolumeSpecName "kube-api-access-85rnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.493597 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" (UID: "e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.575628 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.575667 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85rnj\" (UniqueName: \"kubernetes.io/projected/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2-kube-api-access-85rnj\") on node \"crc\" DevicePath \"\"" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.953949 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl67" event={"ID":"e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2","Type":"ContainerDied","Data":"2703e5b398bfb21f10bb13b2d58facdcf2d9b95ebc8d72f2705273c860c36695"} Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.954025 4758 scope.go:117] "RemoveContainer" containerID="0be6a4a0f5f48c33ab9173badcabd6c2822de5d3093ff7bf485abf420ddfb4fa" Dec 03 18:43:53 crc kubenswrapper[4758]: I1203 18:43:53.954045 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl67" Dec 03 18:43:54 crc kubenswrapper[4758]: I1203 18:43:54.028697 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl67"] Dec 03 18:43:54 crc kubenswrapper[4758]: I1203 18:43:54.057017 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl67"] Dec 03 18:43:54 crc kubenswrapper[4758]: I1203 18:43:54.870149 4758 scope.go:117] "RemoveContainer" containerID="5c340af146e477c77044436457a21f57c92f51f43d0f32097dd7cedaddb750b7" Dec 03 18:43:54 crc kubenswrapper[4758]: I1203 18:43:54.910565 4758 scope.go:117] "RemoveContainer" containerID="cd75fc4c6dee63f24256f4ddac9370009ab5dcc7690b00c04308cae27e85f227" Dec 03 18:43:55 crc kubenswrapper[4758]: I1203 18:43:55.130872 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" path="/var/lib/kubelet/pods/e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2/volumes" Dec 03 18:43:55 crc kubenswrapper[4758]: I1203 18:43:55.988271 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" event={"ID":"373f7373-1c0a-4261-81b5-03a8efcf147e","Type":"ContainerStarted","Data":"9e0d1e105e139c361ecfe236859b34878971f4b966c57a4924003b9965e67793"} Dec 03 18:43:55 crc kubenswrapper[4758]: I1203 18:43:55.992238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-h79l8" event={"ID":"424f6469-230e-4f6b-b257-7ad7feecd589","Type":"ContainerStarted","Data":"02d1c9a44bff6f59a7802e50703aa4bb7713827114edcdd4e572438dd25912b3"} Dec 03 18:43:55 crc kubenswrapper[4758]: I1203 18:43:55.994111 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:43:55 crc kubenswrapper[4758]: I1203 18:43:55.995816 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" event={"ID":"18a96e71-fd5d-4816-a001-92ae35d94c17","Type":"ContainerStarted","Data":"cbcf5caec8bf49bf5c7a447217167f40a3364b4a64d2f9f5b90d51052ad7e102"} Dec 03 18:43:55 crc kubenswrapper[4758]: I1203 18:43:55.997371 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" event={"ID":"aa3f69cb-5012-4879-864e-5826e71f2aa8","Type":"ContainerStarted","Data":"862a380efd65ed595d142ed6823e7b6cf93766b8eee70affdb90e15170397e18"} Dec 03 18:43:56 crc kubenswrapper[4758]: I1203 18:43:56.021032 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rkkv8" podStartSLOduration=3.014975039 podStartE2EDuration="9.021015985s" podCreationTimestamp="2025-12-03 18:43:47 +0000 UTC" firstStartedPulling="2025-12-03 18:43:48.905629659 +0000 UTC m=+6484.107006520" lastFinishedPulling="2025-12-03 18:43:54.911670605 +0000 UTC m=+6490.113047466" observedRunningTime="2025-12-03 18:43:56.013798951 +0000 UTC m=+6491.215175842" watchObservedRunningTime="2025-12-03 18:43:56.021015985 +0000 UTC m=+6491.222392846" Dec 03 18:43:56 crc kubenswrapper[4758]: I1203 18:43:56.045850 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b" podStartSLOduration=3.104192634 podStartE2EDuration="9.045823034s" podCreationTimestamp="2025-12-03 18:43:47 +0000 UTC" firstStartedPulling="2025-12-03 18:43:48.972273646 +0000 UTC m=+6484.173650507" lastFinishedPulling="2025-12-03 18:43:54.913904046 +0000 UTC m=+6490.115280907" observedRunningTime="2025-12-03 18:43:56.029505735 +0000 UTC m=+6491.230882596" watchObservedRunningTime="2025-12-03 18:43:56.045823034 +0000 UTC m=+6491.247199895" Dec 03 18:43:56 crc kubenswrapper[4758]: I1203 18:43:56.066259 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6976796f4b-tgltn" podStartSLOduration=3.268713772 podStartE2EDuration="9.066238915s" podCreationTimestamp="2025-12-03 18:43:47 +0000 UTC" firstStartedPulling="2025-12-03 18:43:49.080808643 +0000 UTC m=+6484.282185504" lastFinishedPulling="2025-12-03 18:43:54.878333786 +0000 UTC m=+6490.079710647" observedRunningTime="2025-12-03 18:43:56.057449408 +0000 UTC m=+6491.258826269" watchObservedRunningTime="2025-12-03 18:43:56.066238915 +0000 UTC m=+6491.267615776" Dec 03 18:43:56 crc kubenswrapper[4758]: I1203 18:43:56.108021 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-h79l8" podStartSLOduration=3.411581314 podStartE2EDuration="9.107999571s" podCreationTimestamp="2025-12-03 18:43:47 +0000 UTC" firstStartedPulling="2025-12-03 18:43:49.194346604 +0000 UTC m=+6484.395723465" lastFinishedPulling="2025-12-03 18:43:54.890764871 +0000 UTC m=+6490.092141722" observedRunningTime="2025-12-03 18:43:56.094104087 +0000 UTC m=+6491.295480968" watchObservedRunningTime="2025-12-03 18:43:56.107999571 +0000 UTC m=+6491.309376432" Dec 03 18:44:02 crc kubenswrapper[4758]: I1203 18:44:02.117760 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" event={"ID":"dd43e0c9-4231-459e-8d32-eddb92573c5b","Type":"ContainerStarted","Data":"9d0c1b9ec67c639795f585be9ff2e1eeb50fbc49ee7b46ee9a4a8ce3c6eab2df"} Dec 03 18:44:02 crc kubenswrapper[4758]: I1203 18:44:02.118890 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:44:02 crc kubenswrapper[4758]: I1203 18:44:02.120345 4758 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-r27q7 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.131:8081/healthz\": dial tcp 10.217.1.131:8081: connect: connection refused" start-of-body= Dec 03 18:44:02 crc kubenswrapper[4758]: I1203 18:44:02.120418 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" podUID="dd43e0c9-4231-459e-8d32-eddb92573c5b" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.131:8081/healthz\": dial tcp 10.217.1.131:8081: connect: connection refused" Dec 03 18:44:02 crc kubenswrapper[4758]: I1203 18:44:02.156723 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" podStartSLOduration=2.579270862 podStartE2EDuration="15.156700368s" podCreationTimestamp="2025-12-03 18:43:47 +0000 UTC" firstStartedPulling="2025-12-03 18:43:49.214917328 +0000 UTC m=+6484.416294189" lastFinishedPulling="2025-12-03 18:44:01.792346834 +0000 UTC m=+6496.993723695" observedRunningTime="2025-12-03 18:44:02.144399086 +0000 UTC m=+6497.345775947" watchObservedRunningTime="2025-12-03 18:44:02.156700368 +0000 UTC m=+6497.358077229" Dec 03 18:44:03 crc kubenswrapper[4758]: I1203 18:44:03.128729 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-r27q7" Dec 03 18:44:08 crc kubenswrapper[4758]: I1203 18:44:08.425233 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-h79l8" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.060116 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-b55f9"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.068049 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-b55f9"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.127891 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c7eeb6e-1676-45f8-9f6d-c704c0be486f" path="/var/lib/kubelet/pods/8c7eeb6e-1676-45f8-9f6d-c704c0be486f/volumes" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.351941 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.352144 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="90b85a0f-5936-4bab-a550-83934e007e9b" containerName="openstackclient" containerID="cri-o://077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" gracePeriod=2 Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.387190 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.403109 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.403174 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.403229 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.403951 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.404020 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" gracePeriod=600 Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427037 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427475 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="extract-utilities" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427496 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="extract-utilities" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427524 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="extract-content" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427533 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="extract-content" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427549 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="registry-server" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427558 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="registry-server" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427575 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="extract-utilities" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427582 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="extract-utilities" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427592 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="registry-server" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427598 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="registry-server" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427606 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90b85a0f-5936-4bab-a550-83934e007e9b" containerName="openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427612 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="90b85a0f-5936-4bab-a550-83934e007e9b" containerName="openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.427631 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="extract-content" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427637 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="extract-content" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427845 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="90b85a0f-5936-4bab-a550-83934e007e9b" containerName="openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427864 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba406b4a-4146-4bf5-ab95-71a881d14143" containerName="registry-server" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.427877 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a36ab6-9ae8-40a3-b73f-7389f5d5b0f2" containerName="registry-server" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.428560 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.443125 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="90b85a0f-5936-4bab-a550-83934e007e9b" podUID="bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.456103 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.479786 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-openstack-config\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.492310 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-openstack-config-secret\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.492548 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfr9g\" (UniqueName: \"kubernetes.io/projected/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-kube-api-access-hfr9g\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.598031 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-openstack-config-secret\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.598570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfr9g\" (UniqueName: \"kubernetes.io/projected/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-kube-api-access-hfr9g\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.598757 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-openstack-config\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.599560 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-openstack-config\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.611608 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-openstack-config-secret\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.650972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfr9g\" (UniqueName: \"kubernetes.io/projected/bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75-kube-api-access-hfr9g\") pod \"openstackclient\" (UID: \"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75\") " pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: E1203 18:44:11.659481 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.720742 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.722228 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.726771 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-wvkt6" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.753064 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.785525 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.803472 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5vtt\" (UniqueName: \"kubernetes.io/projected/07dd279b-b4fa-4544-9afe-1e0639826a54-kube-api-access-v5vtt\") pod \"kube-state-metrics-0\" (UID: \"07dd279b-b4fa-4544-9afe-1e0639826a54\") " pod="openstack/kube-state-metrics-0" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.905469 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5vtt\" (UniqueName: \"kubernetes.io/projected/07dd279b-b4fa-4544-9afe-1e0639826a54-kube-api-access-v5vtt\") pod \"kube-state-metrics-0\" (UID: \"07dd279b-b4fa-4544-9afe-1e0639826a54\") " pod="openstack/kube-state-metrics-0" Dec 03 18:44:11 crc kubenswrapper[4758]: I1203 18:44:11.933933 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5vtt\" (UniqueName: \"kubernetes.io/projected/07dd279b-b4fa-4544-9afe-1e0639826a54-kube-api-access-v5vtt\") pod \"kube-state-metrics-0\" (UID: \"07dd279b-b4fa-4544-9afe-1e0639826a54\") " pod="openstack/kube-state-metrics-0" Dec 03 18:44:12 crc kubenswrapper[4758]: I1203 18:44:12.087330 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 18:44:12 crc kubenswrapper[4758]: I1203 18:44:12.331934 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" exitCode=0 Dec 03 18:44:12 crc kubenswrapper[4758]: I1203 18:44:12.331985 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00"} Dec 03 18:44:12 crc kubenswrapper[4758]: I1203 18:44:12.332021 4758 scope.go:117] "RemoveContainer" containerID="55064fb8d522abc0cbfb578b1ed4ff679982f3cf1154c3a5e9439b6fbc880f07" Dec 03 18:44:12 crc kubenswrapper[4758]: I1203 18:44:12.332865 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:44:12 crc kubenswrapper[4758]: E1203 18:44:12.333322 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:44:12 crc kubenswrapper[4758]: I1203 18:44:12.592099 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 18:44:13 crc kubenswrapper[4758]: I1203 18:44:13.248655 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 18:44:13 crc kubenswrapper[4758]: W1203 18:44:13.248805 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07dd279b_b4fa_4544_9afe_1e0639826a54.slice/crio-8fc659bafc875c9e37256e9fddde0101be5bdf435097e823955297ef6db5093c WatchSource:0}: Error finding container 8fc659bafc875c9e37256e9fddde0101be5bdf435097e823955297ef6db5093c: Status 404 returned error can't find the container with id 8fc659bafc875c9e37256e9fddde0101be5bdf435097e823955297ef6db5093c Dec 03 18:44:13 crc kubenswrapper[4758]: I1203 18:44:13.353573 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"07dd279b-b4fa-4544-9afe-1e0639826a54","Type":"ContainerStarted","Data":"8fc659bafc875c9e37256e9fddde0101be5bdf435097e823955297ef6db5093c"} Dec 03 18:44:13 crc kubenswrapper[4758]: I1203 18:44:13.355524 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75","Type":"ContainerStarted","Data":"00f6e7f1a704705834f112819f18e26a7b03912100f216621e65f8310fbb9f57"} Dec 03 18:44:13 crc kubenswrapper[4758]: I1203 18:44:13.355560 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75","Type":"ContainerStarted","Data":"0c969f4135f5cb360dd9f368f9dc031dea9da723c70b44ddfbc586c84c899fb6"} Dec 03 18:44:13 crc kubenswrapper[4758]: I1203 18:44:13.386825 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.386789025 podStartE2EDuration="2.386789025s" podCreationTimestamp="2025-12-03 18:44:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:44:13.370863366 +0000 UTC m=+6508.572240227" watchObservedRunningTime="2025-12-03 18:44:13.386789025 +0000 UTC m=+6508.588165886" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.011579 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.077216 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config-secret\") pod \"90b85a0f-5936-4bab-a550-83934e007e9b\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.077370 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config\") pod \"90b85a0f-5936-4bab-a550-83934e007e9b\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.077427 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98ct4\" (UniqueName: \"kubernetes.io/projected/90b85a0f-5936-4bab-a550-83934e007e9b-kube-api-access-98ct4\") pod \"90b85a0f-5936-4bab-a550-83934e007e9b\" (UID: \"90b85a0f-5936-4bab-a550-83934e007e9b\") " Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.092984 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90b85a0f-5936-4bab-a550-83934e007e9b-kube-api-access-98ct4" (OuterVolumeSpecName: "kube-api-access-98ct4") pod "90b85a0f-5936-4bab-a550-83934e007e9b" (UID: "90b85a0f-5936-4bab-a550-83934e007e9b"). InnerVolumeSpecName "kube-api-access-98ct4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.109881 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "90b85a0f-5936-4bab-a550-83934e007e9b" (UID: "90b85a0f-5936-4bab-a550-83934e007e9b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.153034 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "90b85a0f-5936-4bab-a550-83934e007e9b" (UID: "90b85a0f-5936-4bab-a550-83934e007e9b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.179719 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.179752 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/90b85a0f-5936-4bab-a550-83934e007e9b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.179761 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98ct4\" (UniqueName: \"kubernetes.io/projected/90b85a0f-5936-4bab-a550-83934e007e9b-kube-api-access-98ct4\") on node \"crc\" DevicePath \"\"" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.330899 4758 scope.go:117] "RemoveContainer" containerID="0a575e5ae3d1e7d4609a1146f1e04466c6e88ef47d66f4fd4e4f690ba0aeadbd" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.376657 4758 generic.go:334] "Generic (PLEG): container finished" podID="90b85a0f-5936-4bab-a550-83934e007e9b" containerID="077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" exitCode=137 Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.376721 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.376781 4758 scope.go:117] "RemoveContainer" containerID="077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.382844 4758 scope.go:117] "RemoveContainer" containerID="9c22843a689b8b4f13ebc902daacc61dc49234d21a90b2c175cb56c270b849e5" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.407178 4758 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="90b85a0f-5936-4bab-a550-83934e007e9b" podUID="bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.419140 4758 scope.go:117] "RemoveContainer" containerID="077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" Dec 03 18:44:14 crc kubenswrapper[4758]: E1203 18:44:14.419534 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6\": container with ID starting with 077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6 not found: ID does not exist" containerID="077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.419584 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6"} err="failed to get container status \"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6\": rpc error: code = NotFound desc = could not find container \"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6\": container with ID starting with 077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6 not found: ID does not exist" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.437602 4758 scope.go:117] "RemoveContainer" containerID="49d92bc662262c70244bde1071dd02148d200c1dbcc16958d2146fdf2b7e9500" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.486093 4758 scope.go:117] "RemoveContainer" containerID="4807c161f542c4d5869e5dda45e7ea582c7482598d1e385ecf6cf07bef9f5b15" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.510825 4758 scope.go:117] "RemoveContainer" containerID="06b63bb05d7325264b426bafee2e2c80bc1d086f57ed26e28882e2b72fe7a416" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.538989 4758 scope.go:117] "RemoveContainer" containerID="4d71f0cfcdd494e80f178e79e515c74c807b82103e6484afeedb305b61a28287" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.624253 4758 scope.go:117] "RemoveContainer" containerID="719ab214c2fe87bdd16e99590cba2e3fe4660c7c3c48eb4a426881a6e76e4dad" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.664245 4758 scope.go:117] "RemoveContainer" containerID="f414296e923694ce76c9c55c98252f197da7407b1e07f8f474297c09a50a99c9" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.713500 4758 scope.go:117] "RemoveContainer" containerID="74db9d938180ef9f316a6f59b59b937aa9b0897f0875a115cc87bfdfa642ffce" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.751632 4758 scope.go:117] "RemoveContainer" containerID="077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" Dec 03 18:44:14 crc kubenswrapper[4758]: E1203 18:44:14.752114 4758 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6\": rpc error: code = NotFound desc = could not find container \"077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6\": container with ID starting with 077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6 not found: ID does not exist" containerID="077fd54272eae526bb37d3ae12b2fb8348e2cc1639c3503bfe0c338aee4574a6" Dec 03 18:44:14 crc kubenswrapper[4758]: I1203 18:44:14.752161 4758 scope.go:117] "RemoveContainer" containerID="2946b52948591e4ccd52bdd8de2e84e01e9e09d32b1a02f966c1306b5bc08bf4" Dec 03 18:44:15 crc kubenswrapper[4758]: I1203 18:44:15.127710 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90b85a0f-5936-4bab-a550-83934e007e9b" path="/var/lib/kubelet/pods/90b85a0f-5936-4bab-a550-83934e007e9b/volumes" Dec 03 18:44:15 crc kubenswrapper[4758]: I1203 18:44:15.386100 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"07dd279b-b4fa-4544-9afe-1e0639826a54","Type":"ContainerStarted","Data":"6ea41d3cb5c9b3cf1cc1d74a43d27b16fa2ce62784a5320463d668a1ccca333b"} Dec 03 18:44:15 crc kubenswrapper[4758]: I1203 18:44:15.386294 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 18:44:15 crc kubenswrapper[4758]: I1203 18:44:15.408180 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.557804378 podStartE2EDuration="4.408163276s" podCreationTimestamp="2025-12-03 18:44:11 +0000 UTC" firstStartedPulling="2025-12-03 18:44:13.251392315 +0000 UTC m=+6508.452769176" lastFinishedPulling="2025-12-03 18:44:14.101751203 +0000 UTC m=+6509.303128074" observedRunningTime="2025-12-03 18:44:15.400579062 +0000 UTC m=+6510.601955943" watchObservedRunningTime="2025-12-03 18:44:15.408163276 +0000 UTC m=+6510.609540137" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.296137 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.299111 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.305495 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.305557 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.306192 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.307311 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.313174 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-bhbrd" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.327495 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.342806 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.342902 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.342954 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0850e686-52c2-42fd-bd78-7c0feef9ac6d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.342987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0850e686-52c2-42fd-bd78-7c0feef9ac6d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.343029 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.343060 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0850e686-52c2-42fd-bd78-7c0feef9ac6d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.343341 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh7mq\" (UniqueName: \"kubernetes.io/projected/0850e686-52c2-42fd-bd78-7c0feef9ac6d-kube-api-access-hh7mq\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.444927 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0850e686-52c2-42fd-bd78-7c0feef9ac6d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.444975 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0850e686-52c2-42fd-bd78-7c0feef9ac6d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.445013 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.445033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0850e686-52c2-42fd-bd78-7c0feef9ac6d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.445108 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh7mq\" (UniqueName: \"kubernetes.io/projected/0850e686-52c2-42fd-bd78-7c0feef9ac6d-kube-api-access-hh7mq\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.445171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.445226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.446450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0850e686-52c2-42fd-bd78-7c0feef9ac6d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.452750 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.459212 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0850e686-52c2-42fd-bd78-7c0feef9ac6d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.459672 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.460172 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0850e686-52c2-42fd-bd78-7c0feef9ac6d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.460397 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0850e686-52c2-42fd-bd78-7c0feef9ac6d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.462539 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh7mq\" (UniqueName: \"kubernetes.io/projected/0850e686-52c2-42fd-bd78-7c0feef9ac6d-kube-api-access-hh7mq\") pod \"alertmanager-metric-storage-0\" (UID: \"0850e686-52c2-42fd-bd78-7c0feef9ac6d\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.624768 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.979829 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.988097 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.991135 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.991250 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.991414 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.991469 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.991591 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-2kjbl" Dec 03 18:44:17 crc kubenswrapper[4758]: I1203 18:44:17.991698 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.020757 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065353 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f07044df-c38c-4897-b412-698ab28b784a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f07044df-c38c-4897-b412-698ab28b784a\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065406 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065432 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065488 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065528 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgjsq\" (UniqueName: \"kubernetes.io/projected/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-kube-api-access-kgjsq\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065592 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-config\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.065665 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.120673 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167255 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-config\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167402 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f07044df-c38c-4897-b412-698ab28b784a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f07044df-c38c-4897-b412-698ab28b784a\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167473 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167494 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167555 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167585 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.167599 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgjsq\" (UniqueName: \"kubernetes.io/projected/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-kube-api-access-kgjsq\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.170210 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.177882 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.178058 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.187102 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.189909 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.190351 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-config\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.201261 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgjsq\" (UniqueName: \"kubernetes.io/projected/475fe5d8-4473-4c4f-a6ca-4f8a60a10060-kube-api-access-kgjsq\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.228936 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.229227 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f07044df-c38c-4897-b412-698ab28b784a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f07044df-c38c-4897-b412-698ab28b784a\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/da6ed5b788e68fabef6c5643fd99cd412ce75b32942b345e116229e269a7a980/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.325237 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f07044df-c38c-4897-b412-698ab28b784a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f07044df-c38c-4897-b412-698ab28b784a\") pod \"prometheus-metric-storage-0\" (UID: \"475fe5d8-4473-4c4f-a6ca-4f8a60a10060\") " pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.413723 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0850e686-52c2-42fd-bd78-7c0feef9ac6d","Type":"ContainerStarted","Data":"eda01c545e9a62babe4cdc0c4992f640e88a23ed3b30f9a7596813d2a7ce2a7e"} Dec 03 18:44:18 crc kubenswrapper[4758]: I1203 18:44:18.619107 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:19 crc kubenswrapper[4758]: I1203 18:44:19.175519 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 18:44:19 crc kubenswrapper[4758]: I1203 18:44:19.424972 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"475fe5d8-4473-4c4f-a6ca-4f8a60a10060","Type":"ContainerStarted","Data":"edf0b4b2a7917e2c0b6c5e4c8a9f94b4c4eaab1f2ed518eba5e140f02418385f"} Dec 03 18:44:22 crc kubenswrapper[4758]: I1203 18:44:22.097190 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 18:44:23 crc kubenswrapper[4758]: I1203 18:44:23.114304 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:44:23 crc kubenswrapper[4758]: E1203 18:44:23.114936 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:44:25 crc kubenswrapper[4758]: I1203 18:44:25.479162 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0850e686-52c2-42fd-bd78-7c0feef9ac6d","Type":"ContainerStarted","Data":"488ff978e483c44cad3c3cb66a16bbec136cb7581fd0e985807afe5f1ac60254"} Dec 03 18:44:25 crc kubenswrapper[4758]: I1203 18:44:25.481283 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"475fe5d8-4473-4c4f-a6ca-4f8a60a10060","Type":"ContainerStarted","Data":"325a3dc2a808d67d7f48490b5760012234ae699721d3397004820e6aac9b92f8"} Dec 03 18:44:30 crc kubenswrapper[4758]: I1203 18:44:30.270927 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-gxqxh" podUID="272d2f0a-3217-4e67-82f3-43dc7aa334f5" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:44:32 crc kubenswrapper[4758]: I1203 18:44:32.554830 4758 generic.go:334] "Generic (PLEG): container finished" podID="0850e686-52c2-42fd-bd78-7c0feef9ac6d" containerID="488ff978e483c44cad3c3cb66a16bbec136cb7581fd0e985807afe5f1ac60254" exitCode=0 Dec 03 18:44:32 crc kubenswrapper[4758]: I1203 18:44:32.554924 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0850e686-52c2-42fd-bd78-7c0feef9ac6d","Type":"ContainerDied","Data":"488ff978e483c44cad3c3cb66a16bbec136cb7581fd0e985807afe5f1ac60254"} Dec 03 18:44:32 crc kubenswrapper[4758]: I1203 18:44:32.557818 4758 generic.go:334] "Generic (PLEG): container finished" podID="475fe5d8-4473-4c4f-a6ca-4f8a60a10060" containerID="325a3dc2a808d67d7f48490b5760012234ae699721d3397004820e6aac9b92f8" exitCode=0 Dec 03 18:44:32 crc kubenswrapper[4758]: I1203 18:44:32.557857 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"475fe5d8-4473-4c4f-a6ca-4f8a60a10060","Type":"ContainerDied","Data":"325a3dc2a808d67d7f48490b5760012234ae699721d3397004820e6aac9b92f8"} Dec 03 18:44:37 crc kubenswrapper[4758]: I1203 18:44:37.115231 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:44:37 crc kubenswrapper[4758]: E1203 18:44:37.116180 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:44:39 crc kubenswrapper[4758]: I1203 18:44:39.643364 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0850e686-52c2-42fd-bd78-7c0feef9ac6d","Type":"ContainerStarted","Data":"706bbc962a428a7525f733f9510ec627b6435ad30af07630fbd5b1be4851cb72"} Dec 03 18:44:39 crc kubenswrapper[4758]: I1203 18:44:39.646043 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"475fe5d8-4473-4c4f-a6ca-4f8a60a10060","Type":"ContainerStarted","Data":"4fd390091f798827329040215d2be81e442561c467f3acf7d9aea3b5ab0db512"} Dec 03 18:44:43 crc kubenswrapper[4758]: I1203 18:44:43.692965 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0850e686-52c2-42fd-bd78-7c0feef9ac6d","Type":"ContainerStarted","Data":"f85e1ea8de976146c3d3330edb04d8fa0a8610b8513acd75d6d67c0779e814b9"} Dec 03 18:44:43 crc kubenswrapper[4758]: I1203 18:44:43.693555 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:43 crc kubenswrapper[4758]: I1203 18:44:43.698063 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 03 18:44:43 crc kubenswrapper[4758]: I1203 18:44:43.698215 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"475fe5d8-4473-4c4f-a6ca-4f8a60a10060","Type":"ContainerStarted","Data":"ce33978b45f93681df86c11182e7461cb3634201b1d5cef65839a6e8bb2f8350"} Dec 03 18:44:43 crc kubenswrapper[4758]: I1203 18:44:43.741094 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.774809278 podStartE2EDuration="26.741050115s" podCreationTimestamp="2025-12-03 18:44:17 +0000 UTC" firstStartedPulling="2025-12-03 18:44:18.135928613 +0000 UTC m=+6513.337305474" lastFinishedPulling="2025-12-03 18:44:39.10216944 +0000 UTC m=+6534.303546311" observedRunningTime="2025-12-03 18:44:43.733217803 +0000 UTC m=+6538.934594674" watchObservedRunningTime="2025-12-03 18:44:43.741050115 +0000 UTC m=+6538.942426986" Dec 03 18:44:46 crc kubenswrapper[4758]: I1203 18:44:46.733209 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"475fe5d8-4473-4c4f-a6ca-4f8a60a10060","Type":"ContainerStarted","Data":"0b0834bad7f3d4ce50bd1dc9716785786dba2f1810f08a0f743bb90ae456c485"} Dec 03 18:44:46 crc kubenswrapper[4758]: I1203 18:44:46.767998 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.76515898 podStartE2EDuration="30.767959217s" podCreationTimestamp="2025-12-03 18:44:16 +0000 UTC" firstStartedPulling="2025-12-03 18:44:19.184183966 +0000 UTC m=+6514.385560827" lastFinishedPulling="2025-12-03 18:44:46.186984213 +0000 UTC m=+6541.388361064" observedRunningTime="2025-12-03 18:44:46.752968963 +0000 UTC m=+6541.954345824" watchObservedRunningTime="2025-12-03 18:44:46.767959217 +0000 UTC m=+6541.969336098" Dec 03 18:44:48 crc kubenswrapper[4758]: I1203 18:44:48.619867 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:48 crc kubenswrapper[4758]: I1203 18:44:48.620360 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:48 crc kubenswrapper[4758]: I1203 18:44:48.623350 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:48 crc kubenswrapper[4758]: I1203 18:44:48.752052 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 18:44:50 crc kubenswrapper[4758]: I1203 18:44:50.114699 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:44:50 crc kubenswrapper[4758]: E1203 18:44:50.115427 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:44:50 crc kubenswrapper[4758]: I1203 18:44:50.921749 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:44:50 crc kubenswrapper[4758]: I1203 18:44:50.935473 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:44:50 crc kubenswrapper[4758]: I1203 18:44:50.935594 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:44:50 crc kubenswrapper[4758]: I1203 18:44:50.949212 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 18:44:50 crc kubenswrapper[4758]: I1203 18:44:50.949590 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.095671 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-config-data\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.095773 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-run-httpd\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.096047 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtkjh\" (UniqueName: \"kubernetes.io/projected/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-kube-api-access-jtkjh\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.096146 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.096236 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.096350 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-log-httpd\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.096457 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-scripts\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198245 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-scripts\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198308 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-config-data\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198345 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-run-httpd\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198487 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtkjh\" (UniqueName: \"kubernetes.io/projected/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-kube-api-access-jtkjh\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198529 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198554 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.198618 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-log-httpd\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.199075 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-log-httpd\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.199994 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-run-httpd\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.210662 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.212751 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-config-data\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.223102 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.232628 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtkjh\" (UniqueName: \"kubernetes.io/projected/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-kube-api-access-jtkjh\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.233527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-scripts\") pod \"ceilometer-0\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.267846 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:44:51 crc kubenswrapper[4758]: I1203 18:44:51.857114 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:44:52 crc kubenswrapper[4758]: I1203 18:44:52.789560 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerStarted","Data":"c7732eca5821f92221c3dcb4ef5a0ef71caecb07c8e9a8e47f6cf11292a45923"} Dec 03 18:44:53 crc kubenswrapper[4758]: I1203 18:44:53.810586 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerStarted","Data":"06a333e0a4cc9b0e56651a8d95d2b5c959b2664b12ee2674f8bdb188833fefdc"} Dec 03 18:44:54 crc kubenswrapper[4758]: I1203 18:44:54.043119 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db92-account-create-update-vgcjb"] Dec 03 18:44:54 crc kubenswrapper[4758]: I1203 18:44:54.059215 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rtkk7"] Dec 03 18:44:54 crc kubenswrapper[4758]: I1203 18:44:54.068411 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db92-account-create-update-vgcjb"] Dec 03 18:44:54 crc kubenswrapper[4758]: I1203 18:44:54.078617 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rtkk7"] Dec 03 18:44:54 crc kubenswrapper[4758]: I1203 18:44:54.820848 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerStarted","Data":"7d77b3d42d8f3cf93ae0b5ac35910e270fe8f4d53260505dd3a0d24dca9ea6ea"} Dec 03 18:44:55 crc kubenswrapper[4758]: I1203 18:44:55.138687 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f4b46e-0c7d-4390-9452-d462f1569eeb" path="/var/lib/kubelet/pods/e5f4b46e-0c7d-4390-9452-d462f1569eeb/volumes" Dec 03 18:44:55 crc kubenswrapper[4758]: I1203 18:44:55.140266 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3f27a8-f563-4c6a-97cc-e595247ea87a" path="/var/lib/kubelet/pods/ff3f27a8-f563-4c6a-97cc-e595247ea87a/volumes" Dec 03 18:44:55 crc kubenswrapper[4758]: I1203 18:44:55.832935 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerStarted","Data":"50bd07e1b5f73678a07b84e71888f7a0a72eab9fb4b76b798058ab3663d64f11"} Dec 03 18:44:56 crc kubenswrapper[4758]: I1203 18:44:56.847275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerStarted","Data":"c1816a1f80f8e6d2e19d1289d7da00d1ba39c4f06394f1dd1acdd429c6bf117d"} Dec 03 18:44:56 crc kubenswrapper[4758]: I1203 18:44:56.847781 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 18:44:56 crc kubenswrapper[4758]: I1203 18:44:56.880249 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.209359109 podStartE2EDuration="6.880228627s" podCreationTimestamp="2025-12-03 18:44:50 +0000 UTC" firstStartedPulling="2025-12-03 18:44:51.849919118 +0000 UTC m=+6547.051295979" lastFinishedPulling="2025-12-03 18:44:56.520788636 +0000 UTC m=+6551.722165497" observedRunningTime="2025-12-03 18:44:56.86736028 +0000 UTC m=+6552.068737161" watchObservedRunningTime="2025-12-03 18:44:56.880228627 +0000 UTC m=+6552.081605488" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.135593 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt"] Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.138061 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.140989 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.140989 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.146193 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt"] Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.312280 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a546bafd-cf70-4fb6-867d-e2dbf24b9064-config-volume\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.312535 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a546bafd-cf70-4fb6-867d-e2dbf24b9064-secret-volume\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.312636 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7gpj\" (UniqueName: \"kubernetes.io/projected/a546bafd-cf70-4fb6-867d-e2dbf24b9064-kube-api-access-l7gpj\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.414312 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a546bafd-cf70-4fb6-867d-e2dbf24b9064-secret-volume\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.414630 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7gpj\" (UniqueName: \"kubernetes.io/projected/a546bafd-cf70-4fb6-867d-e2dbf24b9064-kube-api-access-l7gpj\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.414798 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a546bafd-cf70-4fb6-867d-e2dbf24b9064-config-volume\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.415613 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a546bafd-cf70-4fb6-867d-e2dbf24b9064-config-volume\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.424402 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a546bafd-cf70-4fb6-867d-e2dbf24b9064-secret-volume\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.437820 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7gpj\" (UniqueName: \"kubernetes.io/projected/a546bafd-cf70-4fb6-867d-e2dbf24b9064-kube-api-access-l7gpj\") pod \"collect-profiles-29413125-h4ztt\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:00 crc kubenswrapper[4758]: I1203 18:45:00.464089 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:01 crc kubenswrapper[4758]: I1203 18:45:01.400957 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt"] Dec 03 18:45:01 crc kubenswrapper[4758]: I1203 18:45:01.897894 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" event={"ID":"a546bafd-cf70-4fb6-867d-e2dbf24b9064","Type":"ContainerStarted","Data":"6e893e2641ea4b1873789ef963dddad2aedf33f375da9cc5eed8eb7c22136578"} Dec 03 18:45:02 crc kubenswrapper[4758]: I1203 18:45:02.033634 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xqv27"] Dec 03 18:45:02 crc kubenswrapper[4758]: I1203 18:45:02.044116 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xqv27"] Dec 03 18:45:02 crc kubenswrapper[4758]: I1203 18:45:02.923971 4758 generic.go:334] "Generic (PLEG): container finished" podID="a546bafd-cf70-4fb6-867d-e2dbf24b9064" containerID="fa63e34d478e1e941af28ff6ccb4398c0b34f316ccf37d25ba6153f4947a32c8" exitCode=0 Dec 03 18:45:02 crc kubenswrapper[4758]: I1203 18:45:02.924052 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" event={"ID":"a546bafd-cf70-4fb6-867d-e2dbf24b9064","Type":"ContainerDied","Data":"fa63e34d478e1e941af28ff6ccb4398c0b34f316ccf37d25ba6153f4947a32c8"} Dec 03 18:45:02 crc kubenswrapper[4758]: I1203 18:45:02.993071 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-m5jpt"] Dec 03 18:45:02 crc kubenswrapper[4758]: I1203 18:45:02.995872 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.014858 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-m5jpt"] Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.073656 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7bn4\" (UniqueName: \"kubernetes.io/projected/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-kube-api-access-k7bn4\") pod \"aodh-db-create-m5jpt\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.073851 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-operator-scripts\") pod \"aodh-db-create-m5jpt\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.129743 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5666241-ad1d-4826-acce-2f154cbb8d6b" path="/var/lib/kubelet/pods/b5666241-ad1d-4826-acce-2f154cbb8d6b/volumes" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.160774 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-d98a-account-create-update-r98bs"] Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.163097 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.168865 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.172273 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-d98a-account-create-update-r98bs"] Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.176692 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7bn4\" (UniqueName: \"kubernetes.io/projected/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-kube-api-access-k7bn4\") pod \"aodh-db-create-m5jpt\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.176966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-operator-scripts\") pod \"aodh-db-create-m5jpt\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.178039 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-operator-scripts\") pod \"aodh-db-create-m5jpt\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.209341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7bn4\" (UniqueName: \"kubernetes.io/projected/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-kube-api-access-k7bn4\") pod \"aodh-db-create-m5jpt\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.278748 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gzqv\" (UniqueName: \"kubernetes.io/projected/f803c48e-75e8-45a6-9d60-28961db451ae-kube-api-access-4gzqv\") pod \"aodh-d98a-account-create-update-r98bs\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.278816 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f803c48e-75e8-45a6-9d60-28961db451ae-operator-scripts\") pod \"aodh-d98a-account-create-update-r98bs\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.320197 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.381325 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gzqv\" (UniqueName: \"kubernetes.io/projected/f803c48e-75e8-45a6-9d60-28961db451ae-kube-api-access-4gzqv\") pod \"aodh-d98a-account-create-update-r98bs\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.381403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f803c48e-75e8-45a6-9d60-28961db451ae-operator-scripts\") pod \"aodh-d98a-account-create-update-r98bs\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.382336 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f803c48e-75e8-45a6-9d60-28961db451ae-operator-scripts\") pod \"aodh-d98a-account-create-update-r98bs\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.403750 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gzqv\" (UniqueName: \"kubernetes.io/projected/f803c48e-75e8-45a6-9d60-28961db451ae-kube-api-access-4gzqv\") pod \"aodh-d98a-account-create-update-r98bs\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.483665 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:03 crc kubenswrapper[4758]: W1203 18:45:03.831124 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f7e58a7_b051_4bd5_a0b5_0cd2fe38ccae.slice/crio-0352ca56b5f79e961863b8c0dfe8b3a37c44294f586adfd4718f9f83b49cc378 WatchSource:0}: Error finding container 0352ca56b5f79e961863b8c0dfe8b3a37c44294f586adfd4718f9f83b49cc378: Status 404 returned error can't find the container with id 0352ca56b5f79e961863b8c0dfe8b3a37c44294f586adfd4718f9f83b49cc378 Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.834655 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-m5jpt"] Dec 03 18:45:03 crc kubenswrapper[4758]: I1203 18:45:03.935902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-m5jpt" event={"ID":"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae","Type":"ContainerStarted","Data":"0352ca56b5f79e961863b8c0dfe8b3a37c44294f586adfd4718f9f83b49cc378"} Dec 03 18:45:04 crc kubenswrapper[4758]: W1203 18:45:04.023575 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf803c48e_75e8_45a6_9d60_28961db451ae.slice/crio-dcd975049da3d346ae3be89cb75d9fdadd7528dda012009f9e0141760b7ba3d7 WatchSource:0}: Error finding container dcd975049da3d346ae3be89cb75d9fdadd7528dda012009f9e0141760b7ba3d7: Status 404 returned error can't find the container with id dcd975049da3d346ae3be89cb75d9fdadd7528dda012009f9e0141760b7ba3d7 Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.025962 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-d98a-account-create-update-r98bs"] Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.118624 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:45:04 crc kubenswrapper[4758]: E1203 18:45:04.119084 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.354780 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.503859 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a546bafd-cf70-4fb6-867d-e2dbf24b9064-config-volume\") pod \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.504089 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7gpj\" (UniqueName: \"kubernetes.io/projected/a546bafd-cf70-4fb6-867d-e2dbf24b9064-kube-api-access-l7gpj\") pod \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.504138 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a546bafd-cf70-4fb6-867d-e2dbf24b9064-secret-volume\") pod \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\" (UID: \"a546bafd-cf70-4fb6-867d-e2dbf24b9064\") " Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.504484 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a546bafd-cf70-4fb6-867d-e2dbf24b9064-config-volume" (OuterVolumeSpecName: "config-volume") pod "a546bafd-cf70-4fb6-867d-e2dbf24b9064" (UID: "a546bafd-cf70-4fb6-867d-e2dbf24b9064"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.504770 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a546bafd-cf70-4fb6-867d-e2dbf24b9064-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.509364 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a546bafd-cf70-4fb6-867d-e2dbf24b9064-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a546bafd-cf70-4fb6-867d-e2dbf24b9064" (UID: "a546bafd-cf70-4fb6-867d-e2dbf24b9064"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.510204 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a546bafd-cf70-4fb6-867d-e2dbf24b9064-kube-api-access-l7gpj" (OuterVolumeSpecName: "kube-api-access-l7gpj") pod "a546bafd-cf70-4fb6-867d-e2dbf24b9064" (UID: "a546bafd-cf70-4fb6-867d-e2dbf24b9064"). InnerVolumeSpecName "kube-api-access-l7gpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.606933 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7gpj\" (UniqueName: \"kubernetes.io/projected/a546bafd-cf70-4fb6-867d-e2dbf24b9064-kube-api-access-l7gpj\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.606970 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a546bafd-cf70-4fb6-867d-e2dbf24b9064-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.945496 4758 generic.go:334] "Generic (PLEG): container finished" podID="9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" containerID="a7fe3871ab507ceceb3f54e5764fd99385d76ba7abdb49c745d88c109c88fb58" exitCode=0 Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.945550 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-m5jpt" event={"ID":"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae","Type":"ContainerDied","Data":"a7fe3871ab507ceceb3f54e5764fd99385d76ba7abdb49c745d88c109c88fb58"} Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.947536 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" event={"ID":"a546bafd-cf70-4fb6-867d-e2dbf24b9064","Type":"ContainerDied","Data":"6e893e2641ea4b1873789ef963dddad2aedf33f375da9cc5eed8eb7c22136578"} Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.947559 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e893e2641ea4b1873789ef963dddad2aedf33f375da9cc5eed8eb7c22136578" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.947578 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt" Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.949396 4758 generic.go:334] "Generic (PLEG): container finished" podID="f803c48e-75e8-45a6-9d60-28961db451ae" containerID="33dd7c7a17f3cc077f4270dce9b552b5ade7f75811bf8f0260cac244d39aa7be" exitCode=0 Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.949422 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d98a-account-create-update-r98bs" event={"ID":"f803c48e-75e8-45a6-9d60-28961db451ae","Type":"ContainerDied","Data":"33dd7c7a17f3cc077f4270dce9b552b5ade7f75811bf8f0260cac244d39aa7be"} Dec 03 18:45:04 crc kubenswrapper[4758]: I1203 18:45:04.949438 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d98a-account-create-update-r98bs" event={"ID":"f803c48e-75e8-45a6-9d60-28961db451ae","Type":"ContainerStarted","Data":"dcd975049da3d346ae3be89cb75d9fdadd7528dda012009f9e0141760b7ba3d7"} Dec 03 18:45:05 crc kubenswrapper[4758]: I1203 18:45:05.421556 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv"] Dec 03 18:45:05 crc kubenswrapper[4758]: I1203 18:45:05.430447 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-bcnvv"] Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.441529 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.449731 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.553368 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7bn4\" (UniqueName: \"kubernetes.io/projected/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-kube-api-access-k7bn4\") pod \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.553741 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gzqv\" (UniqueName: \"kubernetes.io/projected/f803c48e-75e8-45a6-9d60-28961db451ae-kube-api-access-4gzqv\") pod \"f803c48e-75e8-45a6-9d60-28961db451ae\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.554003 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f803c48e-75e8-45a6-9d60-28961db451ae-operator-scripts\") pod \"f803c48e-75e8-45a6-9d60-28961db451ae\" (UID: \"f803c48e-75e8-45a6-9d60-28961db451ae\") " Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.554054 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-operator-scripts\") pod \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\" (UID: \"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae\") " Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.554778 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f803c48e-75e8-45a6-9d60-28961db451ae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f803c48e-75e8-45a6-9d60-28961db451ae" (UID: "f803c48e-75e8-45a6-9d60-28961db451ae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.554739 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" (UID: "9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.559088 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-kube-api-access-k7bn4" (OuterVolumeSpecName: "kube-api-access-k7bn4") pod "9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" (UID: "9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae"). InnerVolumeSpecName "kube-api-access-k7bn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.559274 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f803c48e-75e8-45a6-9d60-28961db451ae-kube-api-access-4gzqv" (OuterVolumeSpecName: "kube-api-access-4gzqv") pod "f803c48e-75e8-45a6-9d60-28961db451ae" (UID: "f803c48e-75e8-45a6-9d60-28961db451ae"). InnerVolumeSpecName "kube-api-access-4gzqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.656704 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f803c48e-75e8-45a6-9d60-28961db451ae-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.656739 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.656749 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7bn4\" (UniqueName: \"kubernetes.io/projected/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae-kube-api-access-k7bn4\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.656762 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gzqv\" (UniqueName: \"kubernetes.io/projected/f803c48e-75e8-45a6-9d60-28961db451ae-kube-api-access-4gzqv\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.975846 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d98a-account-create-update-r98bs" event={"ID":"f803c48e-75e8-45a6-9d60-28961db451ae","Type":"ContainerDied","Data":"dcd975049da3d346ae3be89cb75d9fdadd7528dda012009f9e0141760b7ba3d7"} Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.975889 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcd975049da3d346ae3be89cb75d9fdadd7528dda012009f9e0141760b7ba3d7" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.975893 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d98a-account-create-update-r98bs" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.977902 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-m5jpt" event={"ID":"9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae","Type":"ContainerDied","Data":"0352ca56b5f79e961863b8c0dfe8b3a37c44294f586adfd4718f9f83b49cc378"} Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.978216 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0352ca56b5f79e961863b8c0dfe8b3a37c44294f586adfd4718f9f83b49cc378" Dec 03 18:45:06 crc kubenswrapper[4758]: I1203 18:45:06.977932 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m5jpt" Dec 03 18:45:07 crc kubenswrapper[4758]: I1203 18:45:07.125972 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5b93102-be10-46be-9cd0-c28fadccc29a" path="/var/lib/kubelet/pods/e5b93102-be10-46be-9cd0-c28fadccc29a/volumes" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.432096 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-nh66c"] Dec 03 18:45:08 crc kubenswrapper[4758]: E1203 18:45:08.432886 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a546bafd-cf70-4fb6-867d-e2dbf24b9064" containerName="collect-profiles" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.432903 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a546bafd-cf70-4fb6-867d-e2dbf24b9064" containerName="collect-profiles" Dec 03 18:45:08 crc kubenswrapper[4758]: E1203 18:45:08.432949 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" containerName="mariadb-database-create" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.432959 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" containerName="mariadb-database-create" Dec 03 18:45:08 crc kubenswrapper[4758]: E1203 18:45:08.432971 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f803c48e-75e8-45a6-9d60-28961db451ae" containerName="mariadb-account-create-update" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.432979 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f803c48e-75e8-45a6-9d60-28961db451ae" containerName="mariadb-account-create-update" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.433217 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f803c48e-75e8-45a6-9d60-28961db451ae" containerName="mariadb-account-create-update" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.433252 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" containerName="mariadb-database-create" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.433268 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a546bafd-cf70-4fb6-867d-e2dbf24b9064" containerName="collect-profiles" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.434244 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.439290 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.439637 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-2vrbs" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.439764 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.439766 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.443349 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-nh66c"] Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.596712 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97nbn\" (UniqueName: \"kubernetes.io/projected/2be426c3-873e-4472-8efc-abd754d826aa-kube-api-access-97nbn\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.596799 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-config-data\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.596845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-combined-ca-bundle\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.596880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-scripts\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.698518 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-scripts\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.698651 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97nbn\" (UniqueName: \"kubernetes.io/projected/2be426c3-873e-4472-8efc-abd754d826aa-kube-api-access-97nbn\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.698732 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-config-data\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.698775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-combined-ca-bundle\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.704428 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-combined-ca-bundle\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.706086 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-scripts\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.710728 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-config-data\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:08 crc kubenswrapper[4758]: I1203 18:45:08.761182 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97nbn\" (UniqueName: \"kubernetes.io/projected/2be426c3-873e-4472-8efc-abd754d826aa-kube-api-access-97nbn\") pod \"aodh-db-sync-nh66c\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:09 crc kubenswrapper[4758]: I1203 18:45:09.057968 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:09 crc kubenswrapper[4758]: I1203 18:45:09.521371 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-nh66c"] Dec 03 18:45:10 crc kubenswrapper[4758]: I1203 18:45:10.024852 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nh66c" event={"ID":"2be426c3-873e-4472-8efc-abd754d826aa","Type":"ContainerStarted","Data":"3ce59d8039ec8ec930d10de3fabbd8fbd0d34f9d6b54f1390c671a9c018d7029"} Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.076033 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nh66c" event={"ID":"2be426c3-873e-4472-8efc-abd754d826aa","Type":"ContainerStarted","Data":"6b459a65bff49fbcef0d38c12fe44a34c36773cafd8dff8903a96ed535ecc7e6"} Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.097064 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-nh66c" podStartSLOduration=2.327439583 podStartE2EDuration="7.097044622s" podCreationTimestamp="2025-12-03 18:45:08 +0000 UTC" firstStartedPulling="2025-12-03 18:45:09.529317524 +0000 UTC m=+6564.730694385" lastFinishedPulling="2025-12-03 18:45:14.298922563 +0000 UTC m=+6569.500299424" observedRunningTime="2025-12-03 18:45:15.092167951 +0000 UTC m=+6570.293544812" watchObservedRunningTime="2025-12-03 18:45:15.097044622 +0000 UTC m=+6570.298421483" Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.112469 4758 scope.go:117] "RemoveContainer" containerID="208dbdd97f246ae0c4ea5387fda7ca671a7a1e806d73c0687a262aa6b5fbeda5" Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.157810 4758 scope.go:117] "RemoveContainer" containerID="e5f4dbed26a69d5eb8c679c05b88617effa0716dd60a8d9592de96c70722c0b3" Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.215406 4758 scope.go:117] "RemoveContainer" containerID="11008d53643cb88d1b75724952e6e3c8cb5ede96b320f8872a83dc341951c585" Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.247876 4758 scope.go:117] "RemoveContainer" containerID="a27b35c5d641bdee69e2177ae789fb9d457bc213b353c95777a7bdbc0b098ee1" Dec 03 18:45:15 crc kubenswrapper[4758]: I1203 18:45:15.293184 4758 scope.go:117] "RemoveContainer" containerID="519680d41f550cf998b5d4e5487c6d0f2d5152a7019fcb8ac684b66b1032ad80" Dec 03 18:45:17 crc kubenswrapper[4758]: I1203 18:45:17.092538 4758 generic.go:334] "Generic (PLEG): container finished" podID="2be426c3-873e-4472-8efc-abd754d826aa" containerID="6b459a65bff49fbcef0d38c12fe44a34c36773cafd8dff8903a96ed535ecc7e6" exitCode=0 Dec 03 18:45:17 crc kubenswrapper[4758]: I1203 18:45:17.092633 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nh66c" event={"ID":"2be426c3-873e-4472-8efc-abd754d826aa","Type":"ContainerDied","Data":"6b459a65bff49fbcef0d38c12fe44a34c36773cafd8dff8903a96ed535ecc7e6"} Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.115721 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:45:18 crc kubenswrapper[4758]: E1203 18:45:18.116478 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.508228 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.631558 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97nbn\" (UniqueName: \"kubernetes.io/projected/2be426c3-873e-4472-8efc-abd754d826aa-kube-api-access-97nbn\") pod \"2be426c3-873e-4472-8efc-abd754d826aa\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.631934 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-combined-ca-bundle\") pod \"2be426c3-873e-4472-8efc-abd754d826aa\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.632003 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-config-data\") pod \"2be426c3-873e-4472-8efc-abd754d826aa\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.632209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-scripts\") pod \"2be426c3-873e-4472-8efc-abd754d826aa\" (UID: \"2be426c3-873e-4472-8efc-abd754d826aa\") " Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.641632 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-scripts" (OuterVolumeSpecName: "scripts") pod "2be426c3-873e-4472-8efc-abd754d826aa" (UID: "2be426c3-873e-4472-8efc-abd754d826aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.646691 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2be426c3-873e-4472-8efc-abd754d826aa-kube-api-access-97nbn" (OuterVolumeSpecName: "kube-api-access-97nbn") pod "2be426c3-873e-4472-8efc-abd754d826aa" (UID: "2be426c3-873e-4472-8efc-abd754d826aa"). InnerVolumeSpecName "kube-api-access-97nbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.668169 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2be426c3-873e-4472-8efc-abd754d826aa" (UID: "2be426c3-873e-4472-8efc-abd754d826aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.672500 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-config-data" (OuterVolumeSpecName: "config-data") pod "2be426c3-873e-4472-8efc-abd754d826aa" (UID: "2be426c3-873e-4472-8efc-abd754d826aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.734749 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.734815 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.734831 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2be426c3-873e-4472-8efc-abd754d826aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:18 crc kubenswrapper[4758]: I1203 18:45:18.734843 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97nbn\" (UniqueName: \"kubernetes.io/projected/2be426c3-873e-4472-8efc-abd754d826aa-kube-api-access-97nbn\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:19 crc kubenswrapper[4758]: I1203 18:45:19.112322 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nh66c" event={"ID":"2be426c3-873e-4472-8efc-abd754d826aa","Type":"ContainerDied","Data":"3ce59d8039ec8ec930d10de3fabbd8fbd0d34f9d6b54f1390c671a9c018d7029"} Dec 03 18:45:19 crc kubenswrapper[4758]: I1203 18:45:19.112363 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ce59d8039ec8ec930d10de3fabbd8fbd0d34f9d6b54f1390c671a9c018d7029" Dec 03 18:45:19 crc kubenswrapper[4758]: I1203 18:45:19.112390 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nh66c" Dec 03 18:45:21 crc kubenswrapper[4758]: I1203 18:45:21.274290 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.018736 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 18:45:23 crc kubenswrapper[4758]: E1203 18:45:23.019337 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be426c3-873e-4472-8efc-abd754d826aa" containerName="aodh-db-sync" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.019349 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be426c3-873e-4472-8efc-abd754d826aa" containerName="aodh-db-sync" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.019570 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2be426c3-873e-4472-8efc-abd754d826aa" containerName="aodh-db-sync" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.025582 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.027222 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-2vrbs" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.028886 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.029084 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.098378 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.123141 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wql7t\" (UniqueName: \"kubernetes.io/projected/ccfcd869-afcd-4a92-a02c-83b861bf54f8-kube-api-access-wql7t\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.123237 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-config-data\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.123393 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.123435 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-scripts\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.225439 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.225515 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-scripts\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.225573 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wql7t\" (UniqueName: \"kubernetes.io/projected/ccfcd869-afcd-4a92-a02c-83b861bf54f8-kube-api-access-wql7t\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.225650 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-config-data\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.232093 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-scripts\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.233632 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-config-data\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.241271 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccfcd869-afcd-4a92-a02c-83b861bf54f8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.245395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wql7t\" (UniqueName: \"kubernetes.io/projected/ccfcd869-afcd-4a92-a02c-83b861bf54f8-kube-api-access-wql7t\") pod \"aodh-0\" (UID: \"ccfcd869-afcd-4a92-a02c-83b861bf54f8\") " pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.364485 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 18:45:23 crc kubenswrapper[4758]: I1203 18:45:23.897200 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 18:45:24 crc kubenswrapper[4758]: I1203 18:45:24.214963 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ccfcd869-afcd-4a92-a02c-83b861bf54f8","Type":"ContainerStarted","Data":"a8d75d573a75e0b930367add23554ea396a51ba38640117565080af9f6ac078d"} Dec 03 18:45:25 crc kubenswrapper[4758]: I1203 18:45:25.211347 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:45:25 crc kubenswrapper[4758]: I1203 18:45:25.211603 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-central-agent" containerID="cri-o://06a333e0a4cc9b0e56651a8d95d2b5c959b2664b12ee2674f8bdb188833fefdc" gracePeriod=30 Dec 03 18:45:25 crc kubenswrapper[4758]: I1203 18:45:25.211628 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="proxy-httpd" containerID="cri-o://c1816a1f80f8e6d2e19d1289d7da00d1ba39c4f06394f1dd1acdd429c6bf117d" gracePeriod=30 Dec 03 18:45:25 crc kubenswrapper[4758]: I1203 18:45:25.211746 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-notification-agent" containerID="cri-o://7d77b3d42d8f3cf93ae0b5ac35910e270fe8f4d53260505dd3a0d24dca9ea6ea" gracePeriod=30 Dec 03 18:45:25 crc kubenswrapper[4758]: I1203 18:45:25.211976 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="sg-core" containerID="cri-o://50bd07e1b5f73678a07b84e71888f7a0a72eab9fb4b76b798058ab3663d64f11" gracePeriod=30 Dec 03 18:45:25 crc kubenswrapper[4758]: I1203 18:45:25.229249 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ccfcd869-afcd-4a92-a02c-83b861bf54f8","Type":"ContainerStarted","Data":"919b24b2328f4725abe8d07895e53e42170169f4557f2b9dcf4563a9291b5723"} Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.242565 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ccfcd869-afcd-4a92-a02c-83b861bf54f8","Type":"ContainerStarted","Data":"ce0250a92b9c4dbb2cfca8de20bb519d24051e3bab8a788b36ba3755a5952767"} Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246206 4758 generic.go:334] "Generic (PLEG): container finished" podID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerID="c1816a1f80f8e6d2e19d1289d7da00d1ba39c4f06394f1dd1acdd429c6bf117d" exitCode=0 Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246237 4758 generic.go:334] "Generic (PLEG): container finished" podID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerID="50bd07e1b5f73678a07b84e71888f7a0a72eab9fb4b76b798058ab3663d64f11" exitCode=2 Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246248 4758 generic.go:334] "Generic (PLEG): container finished" podID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerID="7d77b3d42d8f3cf93ae0b5ac35910e270fe8f4d53260505dd3a0d24dca9ea6ea" exitCode=0 Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246258 4758 generic.go:334] "Generic (PLEG): container finished" podID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerID="06a333e0a4cc9b0e56651a8d95d2b5c959b2664b12ee2674f8bdb188833fefdc" exitCode=0 Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246282 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerDied","Data":"c1816a1f80f8e6d2e19d1289d7da00d1ba39c4f06394f1dd1acdd429c6bf117d"} Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246314 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerDied","Data":"50bd07e1b5f73678a07b84e71888f7a0a72eab9fb4b76b798058ab3663d64f11"} Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246328 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerDied","Data":"7d77b3d42d8f3cf93ae0b5ac35910e270fe8f4d53260505dd3a0d24dca9ea6ea"} Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.246340 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerDied","Data":"06a333e0a4cc9b0e56651a8d95d2b5c959b2664b12ee2674f8bdb188833fefdc"} Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.594404 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707355 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-run-httpd\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-scripts\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707476 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtkjh\" (UniqueName: \"kubernetes.io/projected/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-kube-api-access-jtkjh\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707501 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-sg-core-conf-yaml\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707543 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-config-data\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-log-httpd\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.707718 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-combined-ca-bundle\") pod \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\" (UID: \"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc\") " Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.708435 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.708821 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.714415 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-kube-api-access-jtkjh" (OuterVolumeSpecName: "kube-api-access-jtkjh") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "kube-api-access-jtkjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.717884 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-scripts" (OuterVolumeSpecName: "scripts") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.758117 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.809787 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.809816 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.809826 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtkjh\" (UniqueName: \"kubernetes.io/projected/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-kube-api-access-jtkjh\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.809837 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.809846 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.849608 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-config-data" (OuterVolumeSpecName: "config-data") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.854429 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" (UID: "7049c3f6-9e3f-4c6c-8a83-017a3124b4dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.912040 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:26 crc kubenswrapper[4758]: I1203 18:45:26.912084 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.261798 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7049c3f6-9e3f-4c6c-8a83-017a3124b4dc","Type":"ContainerDied","Data":"c7732eca5821f92221c3dcb4ef5a0ef71caecb07c8e9a8e47f6cf11292a45923"} Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.261847 4758 scope.go:117] "RemoveContainer" containerID="c1816a1f80f8e6d2e19d1289d7da00d1ba39c4f06394f1dd1acdd429c6bf117d" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.261865 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.288798 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.289957 4758 scope.go:117] "RemoveContainer" containerID="50bd07e1b5f73678a07b84e71888f7a0a72eab9fb4b76b798058ab3663d64f11" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.301720 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.313911 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:45:27 crc kubenswrapper[4758]: E1203 18:45:27.314630 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-notification-agent" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.314652 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-notification-agent" Dec 03 18:45:27 crc kubenswrapper[4758]: E1203 18:45:27.314700 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="proxy-httpd" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.314711 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="proxy-httpd" Dec 03 18:45:27 crc kubenswrapper[4758]: E1203 18:45:27.314739 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="sg-core" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.314747 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="sg-core" Dec 03 18:45:27 crc kubenswrapper[4758]: E1203 18:45:27.314778 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-central-agent" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.314787 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-central-agent" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.315040 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-central-agent" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.315064 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="ceilometer-notification-agent" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.315090 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="sg-core" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.315104 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" containerName="proxy-httpd" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.317143 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.319259 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.322292 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.330077 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.352263 4758 scope.go:117] "RemoveContainer" containerID="7d77b3d42d8f3cf93ae0b5ac35910e270fe8f4d53260505dd3a0d24dca9ea6ea" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.391278 4758 scope.go:117] "RemoveContainer" containerID="06a333e0a4cc9b0e56651a8d95d2b5c959b2664b12ee2674f8bdb188833fefdc" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422374 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422490 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-run-httpd\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422529 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd4r8\" (UniqueName: \"kubernetes.io/projected/8a99a1df-a9fb-4906-bc9d-40a814eec360-kube-api-access-vd4r8\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422621 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-config-data\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422664 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-log-httpd\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422729 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.422910 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-scripts\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524539 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-config-data\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524622 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-log-httpd\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524734 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524783 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-scripts\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524829 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524909 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-run-httpd\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.524935 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd4r8\" (UniqueName: \"kubernetes.io/projected/8a99a1df-a9fb-4906-bc9d-40a814eec360-kube-api-access-vd4r8\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.525447 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-run-httpd\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.525518 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-log-httpd\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.529116 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-config-data\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.530480 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.533359 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.533426 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-scripts\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.547569 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd4r8\" (UniqueName: \"kubernetes.io/projected/8a99a1df-a9fb-4906-bc9d-40a814eec360-kube-api-access-vd4r8\") pod \"ceilometer-0\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " pod="openstack/ceilometer-0" Dec 03 18:45:27 crc kubenswrapper[4758]: I1203 18:45:27.637426 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:45:28 crc kubenswrapper[4758]: I1203 18:45:28.119818 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:45:28 crc kubenswrapper[4758]: W1203 18:45:28.127170 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a99a1df_a9fb_4906_bc9d_40a814eec360.slice/crio-faa1ae522e67c0c0a079ac85d80d3d5e571f8999994e3ab50ea1b24e365cba06 WatchSource:0}: Error finding container faa1ae522e67c0c0a079ac85d80d3d5e571f8999994e3ab50ea1b24e365cba06: Status 404 returned error can't find the container with id faa1ae522e67c0c0a079ac85d80d3d5e571f8999994e3ab50ea1b24e365cba06 Dec 03 18:45:28 crc kubenswrapper[4758]: I1203 18:45:28.271302 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ccfcd869-afcd-4a92-a02c-83b861bf54f8","Type":"ContainerStarted","Data":"828687af0f185d0a43bf5ffd737eaf56f2e6ec5e2681d545376c0c757e1ed3e8"} Dec 03 18:45:28 crc kubenswrapper[4758]: I1203 18:45:28.273049 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerStarted","Data":"faa1ae522e67c0c0a079ac85d80d3d5e571f8999994e3ab50ea1b24e365cba06"} Dec 03 18:45:29 crc kubenswrapper[4758]: I1203 18:45:29.125815 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7049c3f6-9e3f-4c6c-8a83-017a3124b4dc" path="/var/lib/kubelet/pods/7049c3f6-9e3f-4c6c-8a83-017a3124b4dc/volumes" Dec 03 18:45:29 crc kubenswrapper[4758]: I1203 18:45:29.283135 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerStarted","Data":"bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331"} Dec 03 18:45:30 crc kubenswrapper[4758]: I1203 18:45:30.294389 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ccfcd869-afcd-4a92-a02c-83b861bf54f8","Type":"ContainerStarted","Data":"f0d59d2c34b61ef82aef3efc3a7b78778490dcd7fce30ffe8c8a1ad719b88095"} Dec 03 18:45:30 crc kubenswrapper[4758]: I1203 18:45:30.296740 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerStarted","Data":"cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24"} Dec 03 18:45:30 crc kubenswrapper[4758]: I1203 18:45:30.380130 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.418274896 podStartE2EDuration="8.38010567s" podCreationTimestamp="2025-12-03 18:45:22 +0000 UTC" firstStartedPulling="2025-12-03 18:45:23.894608187 +0000 UTC m=+6579.095985048" lastFinishedPulling="2025-12-03 18:45:29.856438961 +0000 UTC m=+6585.057815822" observedRunningTime="2025-12-03 18:45:30.374061047 +0000 UTC m=+6585.575437908" watchObservedRunningTime="2025-12-03 18:45:30.38010567 +0000 UTC m=+6585.581482531" Dec 03 18:45:31 crc kubenswrapper[4758]: I1203 18:45:31.310538 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerStarted","Data":"ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27"} Dec 03 18:45:32 crc kubenswrapper[4758]: I1203 18:45:32.114493 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:45:32 crc kubenswrapper[4758]: E1203 18:45:32.114777 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:45:33 crc kubenswrapper[4758]: I1203 18:45:33.332062 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerStarted","Data":"4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21"} Dec 03 18:45:33 crc kubenswrapper[4758]: I1203 18:45:33.332487 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 18:45:33 crc kubenswrapper[4758]: I1203 18:45:33.362567 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.818615779 podStartE2EDuration="6.362547613s" podCreationTimestamp="2025-12-03 18:45:27 +0000 UTC" firstStartedPulling="2025-12-03 18:45:28.130455744 +0000 UTC m=+6583.331832605" lastFinishedPulling="2025-12-03 18:45:32.674387578 +0000 UTC m=+6587.875764439" observedRunningTime="2025-12-03 18:45:33.352966945 +0000 UTC m=+6588.554343826" watchObservedRunningTime="2025-12-03 18:45:33.362547613 +0000 UTC m=+6588.563924474" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.601979 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-5fqlp"] Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.603516 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.612549 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-5fqlp"] Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.708777 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-bc50-account-create-update-gmnkc"] Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.710208 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.712224 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.734740 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-bc50-account-create-update-gmnkc"] Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.750605 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b94047-0675-47ef-bd70-bffe0d411da6-operator-scripts\") pod \"manila-db-create-5fqlp\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.750724 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54dd2\" (UniqueName: \"kubernetes.io/projected/62b94047-0675-47ef-bd70-bffe0d411da6-kube-api-access-54dd2\") pod \"manila-db-create-5fqlp\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.852405 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f6f035-7ce0-458e-ac7d-10767462400d-operator-scripts\") pod \"manila-bc50-account-create-update-gmnkc\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.852463 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b94047-0675-47ef-bd70-bffe0d411da6-operator-scripts\") pod \"manila-db-create-5fqlp\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.852543 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbjc9\" (UniqueName: \"kubernetes.io/projected/99f6f035-7ce0-458e-ac7d-10767462400d-kube-api-access-nbjc9\") pod \"manila-bc50-account-create-update-gmnkc\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.852565 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54dd2\" (UniqueName: \"kubernetes.io/projected/62b94047-0675-47ef-bd70-bffe0d411da6-kube-api-access-54dd2\") pod \"manila-db-create-5fqlp\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.854198 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b94047-0675-47ef-bd70-bffe0d411da6-operator-scripts\") pod \"manila-db-create-5fqlp\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.882541 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54dd2\" (UniqueName: \"kubernetes.io/projected/62b94047-0675-47ef-bd70-bffe0d411da6-kube-api-access-54dd2\") pod \"manila-db-create-5fqlp\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.923005 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.954276 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f6f035-7ce0-458e-ac7d-10767462400d-operator-scripts\") pod \"manila-bc50-account-create-update-gmnkc\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.954366 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbjc9\" (UniqueName: \"kubernetes.io/projected/99f6f035-7ce0-458e-ac7d-10767462400d-kube-api-access-nbjc9\") pod \"manila-bc50-account-create-update-gmnkc\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.955218 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f6f035-7ce0-458e-ac7d-10767462400d-operator-scripts\") pod \"manila-bc50-account-create-update-gmnkc\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:35 crc kubenswrapper[4758]: I1203 18:45:35.970771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbjc9\" (UniqueName: \"kubernetes.io/projected/99f6f035-7ce0-458e-ac7d-10767462400d-kube-api-access-nbjc9\") pod \"manila-bc50-account-create-update-gmnkc\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:36 crc kubenswrapper[4758]: I1203 18:45:36.037308 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:36 crc kubenswrapper[4758]: W1203 18:45:36.578814 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62b94047_0675_47ef_bd70_bffe0d411da6.slice/crio-6ea0717e0c3de4967c14147cd66524304621df11d533262a9edaa5b93c899f41 WatchSource:0}: Error finding container 6ea0717e0c3de4967c14147cd66524304621df11d533262a9edaa5b93c899f41: Status 404 returned error can't find the container with id 6ea0717e0c3de4967c14147cd66524304621df11d533262a9edaa5b93c899f41 Dec 03 18:45:36 crc kubenswrapper[4758]: I1203 18:45:36.580831 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-5fqlp"] Dec 03 18:45:36 crc kubenswrapper[4758]: I1203 18:45:36.685761 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-bc50-account-create-update-gmnkc"] Dec 03 18:45:37 crc kubenswrapper[4758]: I1203 18:45:37.403068 4758 generic.go:334] "Generic (PLEG): container finished" podID="62b94047-0675-47ef-bd70-bffe0d411da6" containerID="0cfbc7b94a728c199a84cff2f2265f566e20f88313dc4dd78cd6404cee1d495f" exitCode=0 Dec 03 18:45:37 crc kubenswrapper[4758]: I1203 18:45:37.403296 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-5fqlp" event={"ID":"62b94047-0675-47ef-bd70-bffe0d411da6","Type":"ContainerDied","Data":"0cfbc7b94a728c199a84cff2f2265f566e20f88313dc4dd78cd6404cee1d495f"} Dec 03 18:45:37 crc kubenswrapper[4758]: I1203 18:45:37.403439 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-5fqlp" event={"ID":"62b94047-0675-47ef-bd70-bffe0d411da6","Type":"ContainerStarted","Data":"6ea0717e0c3de4967c14147cd66524304621df11d533262a9edaa5b93c899f41"} Dec 03 18:45:37 crc kubenswrapper[4758]: I1203 18:45:37.405650 4758 generic.go:334] "Generic (PLEG): container finished" podID="99f6f035-7ce0-458e-ac7d-10767462400d" containerID="7580f0ffaeb2004e4cf8e0068e133822d1042b9b129f2a507420abb145b49524" exitCode=0 Dec 03 18:45:37 crc kubenswrapper[4758]: I1203 18:45:37.405945 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-bc50-account-create-update-gmnkc" event={"ID":"99f6f035-7ce0-458e-ac7d-10767462400d","Type":"ContainerDied","Data":"7580f0ffaeb2004e4cf8e0068e133822d1042b9b129f2a507420abb145b49524"} Dec 03 18:45:37 crc kubenswrapper[4758]: I1203 18:45:37.405989 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-bc50-account-create-update-gmnkc" event={"ID":"99f6f035-7ce0-458e-ac7d-10767462400d","Type":"ContainerStarted","Data":"5d5264f71f0b81f88e1b8e2f49bdfdfb552cf9d1314645c8870d414a2da080f3"} Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.003213 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.010310 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.166853 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54dd2\" (UniqueName: \"kubernetes.io/projected/62b94047-0675-47ef-bd70-bffe0d411da6-kube-api-access-54dd2\") pod \"62b94047-0675-47ef-bd70-bffe0d411da6\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.167200 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbjc9\" (UniqueName: \"kubernetes.io/projected/99f6f035-7ce0-458e-ac7d-10767462400d-kube-api-access-nbjc9\") pod \"99f6f035-7ce0-458e-ac7d-10767462400d\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.167298 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b94047-0675-47ef-bd70-bffe0d411da6-operator-scripts\") pod \"62b94047-0675-47ef-bd70-bffe0d411da6\" (UID: \"62b94047-0675-47ef-bd70-bffe0d411da6\") " Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.167577 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f6f035-7ce0-458e-ac7d-10767462400d-operator-scripts\") pod \"99f6f035-7ce0-458e-ac7d-10767462400d\" (UID: \"99f6f035-7ce0-458e-ac7d-10767462400d\") " Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.168032 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62b94047-0675-47ef-bd70-bffe0d411da6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62b94047-0675-47ef-bd70-bffe0d411da6" (UID: "62b94047-0675-47ef-bd70-bffe0d411da6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.168108 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99f6f035-7ce0-458e-ac7d-10767462400d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99f6f035-7ce0-458e-ac7d-10767462400d" (UID: "99f6f035-7ce0-458e-ac7d-10767462400d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.168748 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b94047-0675-47ef-bd70-bffe0d411da6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.168783 4758 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f6f035-7ce0-458e-ac7d-10767462400d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.173133 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f6f035-7ce0-458e-ac7d-10767462400d-kube-api-access-nbjc9" (OuterVolumeSpecName: "kube-api-access-nbjc9") pod "99f6f035-7ce0-458e-ac7d-10767462400d" (UID: "99f6f035-7ce0-458e-ac7d-10767462400d"). InnerVolumeSpecName "kube-api-access-nbjc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.174314 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b94047-0675-47ef-bd70-bffe0d411da6-kube-api-access-54dd2" (OuterVolumeSpecName: "kube-api-access-54dd2") pod "62b94047-0675-47ef-bd70-bffe0d411da6" (UID: "62b94047-0675-47ef-bd70-bffe0d411da6"). InnerVolumeSpecName "kube-api-access-54dd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.274246 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbjc9\" (UniqueName: \"kubernetes.io/projected/99f6f035-7ce0-458e-ac7d-10767462400d-kube-api-access-nbjc9\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.274278 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54dd2\" (UniqueName: \"kubernetes.io/projected/62b94047-0675-47ef-bd70-bffe0d411da6-kube-api-access-54dd2\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.423655 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-5fqlp" event={"ID":"62b94047-0675-47ef-bd70-bffe0d411da6","Type":"ContainerDied","Data":"6ea0717e0c3de4967c14147cd66524304621df11d533262a9edaa5b93c899f41"} Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.423936 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ea0717e0c3de4967c14147cd66524304621df11d533262a9edaa5b93c899f41" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.423739 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-5fqlp" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.425178 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-bc50-account-create-update-gmnkc" event={"ID":"99f6f035-7ce0-458e-ac7d-10767462400d","Type":"ContainerDied","Data":"5d5264f71f0b81f88e1b8e2f49bdfdfb552cf9d1314645c8870d414a2da080f3"} Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.425208 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d5264f71f0b81f88e1b8e2f49bdfdfb552cf9d1314645c8870d414a2da080f3" Dec 03 18:45:39 crc kubenswrapper[4758]: I1203 18:45:39.425266 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bc50-account-create-update-gmnkc" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.040646 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-2bngz"] Dec 03 18:45:41 crc kubenswrapper[4758]: E1203 18:45:41.041940 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f6f035-7ce0-458e-ac7d-10767462400d" containerName="mariadb-account-create-update" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.041965 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f6f035-7ce0-458e-ac7d-10767462400d" containerName="mariadb-account-create-update" Dec 03 18:45:41 crc kubenswrapper[4758]: E1203 18:45:41.041998 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b94047-0675-47ef-bd70-bffe0d411da6" containerName="mariadb-database-create" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.042007 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b94047-0675-47ef-bd70-bffe0d411da6" containerName="mariadb-database-create" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.042284 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f6f035-7ce0-458e-ac7d-10767462400d" containerName="mariadb-account-create-update" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.042318 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b94047-0675-47ef-bd70-bffe0d411da6" containerName="mariadb-database-create" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.043447 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.048992 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.049285 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-mh28x" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.099780 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-2bngz"] Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.214446 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-config-data\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.214700 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-combined-ca-bundle\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.214824 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5c6l\" (UniqueName: \"kubernetes.io/projected/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-kube-api-access-r5c6l\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.214860 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-job-config-data\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.316491 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-config-data\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.316607 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-combined-ca-bundle\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.316656 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5c6l\" (UniqueName: \"kubernetes.io/projected/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-kube-api-access-r5c6l\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.316687 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-job-config-data\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.322477 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-combined-ca-bundle\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.322584 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-config-data\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.330095 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-job-config-data\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.345160 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5c6l\" (UniqueName: \"kubernetes.io/projected/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-kube-api-access-r5c6l\") pod \"manila-db-sync-2bngz\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.375853 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:41 crc kubenswrapper[4758]: I1203 18:45:41.984267 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-2bngz"] Dec 03 18:45:41 crc kubenswrapper[4758]: W1203 18:45:41.987549 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a2b2cf3_4c51_40d3_a274_01f3cb9a73b0.slice/crio-0bfba73a53ea1865424411a05396fc8076a2941af570d152f55b179e7cb205c6 WatchSource:0}: Error finding container 0bfba73a53ea1865424411a05396fc8076a2941af570d152f55b179e7cb205c6: Status 404 returned error can't find the container with id 0bfba73a53ea1865424411a05396fc8076a2941af570d152f55b179e7cb205c6 Dec 03 18:45:42 crc kubenswrapper[4758]: I1203 18:45:42.474636 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2bngz" event={"ID":"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0","Type":"ContainerStarted","Data":"0bfba73a53ea1865424411a05396fc8076a2941af570d152f55b179e7cb205c6"} Dec 03 18:45:44 crc kubenswrapper[4758]: I1203 18:45:44.114324 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:45:44 crc kubenswrapper[4758]: E1203 18:45:44.115076 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:45:49 crc kubenswrapper[4758]: I1203 18:45:49.548305 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2bngz" event={"ID":"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0","Type":"ContainerStarted","Data":"23a1c883b688c374f4c94115237ebf5d253db8254f8c276989a615a0d01496d7"} Dec 03 18:45:49 crc kubenswrapper[4758]: I1203 18:45:49.564289 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-2bngz" podStartSLOduration=2.416282673 podStartE2EDuration="8.564273767s" podCreationTimestamp="2025-12-03 18:45:41 +0000 UTC" firstStartedPulling="2025-12-03 18:45:41.989866205 +0000 UTC m=+6597.191243066" lastFinishedPulling="2025-12-03 18:45:48.137857299 +0000 UTC m=+6603.339234160" observedRunningTime="2025-12-03 18:45:49.563288611 +0000 UTC m=+6604.764665472" watchObservedRunningTime="2025-12-03 18:45:49.564273767 +0000 UTC m=+6604.765650628" Dec 03 18:45:51 crc kubenswrapper[4758]: I1203 18:45:51.567480 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" containerID="23a1c883b688c374f4c94115237ebf5d253db8254f8c276989a615a0d01496d7" exitCode=0 Dec 03 18:45:51 crc kubenswrapper[4758]: I1203 18:45:51.567580 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2bngz" event={"ID":"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0","Type":"ContainerDied","Data":"23a1c883b688c374f4c94115237ebf5d253db8254f8c276989a615a0d01496d7"} Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.035461 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.161712 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5c6l\" (UniqueName: \"kubernetes.io/projected/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-kube-api-access-r5c6l\") pod \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.161799 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-job-config-data\") pod \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.161860 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-combined-ca-bundle\") pod \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.161926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-config-data\") pod \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\" (UID: \"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0\") " Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.167332 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-kube-api-access-r5c6l" (OuterVolumeSpecName: "kube-api-access-r5c6l") pod "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" (UID: "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0"). InnerVolumeSpecName "kube-api-access-r5c6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.168664 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" (UID: "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.170395 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-config-data" (OuterVolumeSpecName: "config-data") pod "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" (UID: "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.214548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" (UID: "3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.265338 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5c6l\" (UniqueName: \"kubernetes.io/projected/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-kube-api-access-r5c6l\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.265390 4758 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.265410 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.265428 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.591310 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2bngz" event={"ID":"3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0","Type":"ContainerDied","Data":"0bfba73a53ea1865424411a05396fc8076a2941af570d152f55b179e7cb205c6"} Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.591367 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bfba73a53ea1865424411a05396fc8076a2941af570d152f55b179e7cb205c6" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.591404 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2bngz" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.815774 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:45:53 crc kubenswrapper[4758]: E1203 18:45:53.816572 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" containerName="manila-db-sync" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.816591 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" containerName="manila-db-sync" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.816797 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" containerName="manila-db-sync" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.817999 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.826341 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.826780 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.827001 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-mh28x" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.827266 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.832086 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.878618 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.878743 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-config-data\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.878819 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-scripts\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.878982 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.879062 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.879198 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrvq8\" (UniqueName: \"kubernetes.io/projected/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-kube-api-access-qrvq8\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.981736 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.983987 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987189 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrvq8\" (UniqueName: \"kubernetes.io/projected/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-kube-api-access-qrvq8\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987271 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987338 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-config-data\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987381 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-scripts\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987481 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987531 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.987694 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:53 crc kubenswrapper[4758]: I1203 18:45:53.989161 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.002067 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.004318 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-config-data\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.005278 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.020569 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrvq8\" (UniqueName: \"kubernetes.io/projected/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-kube-api-access-qrvq8\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.023503 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d-scripts\") pod \"manila-scheduler-0\" (UID: \"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d\") " pod="openstack/manila-scheduler-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.064761 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bdc5b97bf-pmfc8"] Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.067010 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095113 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095198 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095232 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095289 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095346 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-config-data\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095447 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr6ms\" (UniqueName: \"kubernetes.io/projected/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-kube-api-access-xr6ms\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095516 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-ceph\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095548 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-scripts\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.095719 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.131448 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bdc5b97bf-pmfc8"] Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.173268 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200068 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200149 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lfwp\" (UniqueName: \"kubernetes.io/projected/5209a976-1aab-4069-b293-086b22734c5e-kube-api-access-2lfwp\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200196 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-config-data\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200270 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200314 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr6ms\" (UniqueName: \"kubernetes.io/projected/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-kube-api-access-xr6ms\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-config\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200393 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200441 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-ceph\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-scripts\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200547 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200586 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-dns-svc\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200646 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.200785 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.207020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-config-data\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.207444 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.220144 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-scripts\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.224067 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.227361 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.239242 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-ceph\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.251527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr6ms\" (UniqueName: \"kubernetes.io/projected/e7f8de8f-60ad-4b7e-92c5-6df43aae3589-kube-api-access-xr6ms\") pod \"manila-share-share1-0\" (UID: \"e7f8de8f-60ad-4b7e-92c5-6df43aae3589\") " pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.303102 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.303215 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-config\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.303237 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.303340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-dns-svc\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.303434 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lfwp\" (UniqueName: \"kubernetes.io/projected/5209a976-1aab-4069-b293-086b22734c5e-kube-api-access-2lfwp\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.304793 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.305456 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-dns-svc\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.306511 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-config\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.316961 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.331431 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lfwp\" (UniqueName: \"kubernetes.io/projected/5209a976-1aab-4069-b293-086b22734c5e-kube-api-access-2lfwp\") pod \"dnsmasq-dns-6bdc5b97bf-pmfc8\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.415732 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.418648 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.421260 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.423873 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.433671 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.457932 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508203 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t6xp\" (UniqueName: \"kubernetes.io/projected/068d058a-40e9-4d0c-bd12-56290e862552-kube-api-access-7t6xp\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508298 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508368 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-config-data-custom\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-scripts\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508524 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-config-data\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508605 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/068d058a-40e9-4d0c-bd12-56290e862552-etc-machine-id\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.508661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d058a-40e9-4d0c-bd12-56290e862552-logs\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.620274 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-scripts\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.620813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-config-data\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.621977 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/068d058a-40e9-4d0c-bd12-56290e862552-etc-machine-id\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.622150 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d058a-40e9-4d0c-bd12-56290e862552-logs\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.622373 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t6xp\" (UniqueName: \"kubernetes.io/projected/068d058a-40e9-4d0c-bd12-56290e862552-kube-api-access-7t6xp\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.622493 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.622653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-config-data-custom\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.624715 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d058a-40e9-4d0c-bd12-56290e862552-logs\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.624791 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/068d058a-40e9-4d0c-bd12-56290e862552-etc-machine-id\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.629983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-config-data\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.645754 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-config-data-custom\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.647913 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-scripts\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.648556 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t6xp\" (UniqueName: \"kubernetes.io/projected/068d058a-40e9-4d0c-bd12-56290e862552-kube-api-access-7t6xp\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.648659 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d058a-40e9-4d0c-bd12-56290e862552-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"068d058a-40e9-4d0c-bd12-56290e862552\") " pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.742770 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:45:54 crc kubenswrapper[4758]: I1203 18:45:54.959261 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.143970 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bdc5b97bf-pmfc8"] Dec 03 18:45:55 crc kubenswrapper[4758]: W1203 18:45:55.172002 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5209a976_1aab_4069_b293_086b22734c5e.slice/crio-edd0ed161a35d66f0bcca28cc36beb93667ae7791259efeeb0834f34b93e2c5e WatchSource:0}: Error finding container edd0ed161a35d66f0bcca28cc36beb93667ae7791259efeeb0834f34b93e2c5e: Status 404 returned error can't find the container with id edd0ed161a35d66f0bcca28cc36beb93667ae7791259efeeb0834f34b93e2c5e Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.299122 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.476502 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:45:55 crc kubenswrapper[4758]: W1203 18:45:55.500353 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod068d058a_40e9_4d0c_bd12_56290e862552.slice/crio-7f123236618fd22b340592661bb818504c62fa04f9b3792800108254adfc4f2d WatchSource:0}: Error finding container 7f123236618fd22b340592661bb818504c62fa04f9b3792800108254adfc4f2d: Status 404 returned error can't find the container with id 7f123236618fd22b340592661bb818504c62fa04f9b3792800108254adfc4f2d Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.681446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"068d058a-40e9-4d0c-bd12-56290e862552","Type":"ContainerStarted","Data":"7f123236618fd22b340592661bb818504c62fa04f9b3792800108254adfc4f2d"} Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.684335 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d","Type":"ContainerStarted","Data":"ae7c0f0b54f7fa59b7fe336e8e0c6ff90070388cdc407d814728e3a85d593d37"} Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.695906 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e7f8de8f-60ad-4b7e-92c5-6df43aae3589","Type":"ContainerStarted","Data":"7d0b8caf97af246db046a049c8b985df0c921e712a1ae52e8ccfcc83f0ae504a"} Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.698276 4758 generic.go:334] "Generic (PLEG): container finished" podID="5209a976-1aab-4069-b293-086b22734c5e" containerID="4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7" exitCode=0 Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.698315 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" event={"ID":"5209a976-1aab-4069-b293-086b22734c5e","Type":"ContainerDied","Data":"4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7"} Dec 03 18:45:55 crc kubenswrapper[4758]: I1203 18:45:55.698341 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" event={"ID":"5209a976-1aab-4069-b293-086b22734c5e","Type":"ContainerStarted","Data":"edd0ed161a35d66f0bcca28cc36beb93667ae7791259efeeb0834f34b93e2c5e"} Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.709054 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"068d058a-40e9-4d0c-bd12-56290e862552","Type":"ContainerStarted","Data":"4339c61b115c89198d1acd7daa4abe21a6bf46f75d2e610a0f2b8ee86cd7f33a"} Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.709622 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"068d058a-40e9-4d0c-bd12-56290e862552","Type":"ContainerStarted","Data":"9a8b63284eefd8d01b556d5aa93be19add0629fb638f65fe68e4c86084b7f00c"} Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.709639 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.714188 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d","Type":"ContainerStarted","Data":"d6d78398f90e38a73eabee58ba6b360f28f66c2557991094051915790f029336"} Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.714219 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d","Type":"ContainerStarted","Data":"379135e5d3029831b2885a1f3a13b3ab8be72b809b6901cdc7a887d55c072fe5"} Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.722967 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" event={"ID":"5209a976-1aab-4069-b293-086b22734c5e","Type":"ContainerStarted","Data":"cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256"} Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.723166 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.736950 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.736927509 podStartE2EDuration="2.736927509s" podCreationTimestamp="2025-12-03 18:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:45:56.728585554 +0000 UTC m=+6611.929962415" watchObservedRunningTime="2025-12-03 18:45:56.736927509 +0000 UTC m=+6611.938304370" Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.772531 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.191771681 podStartE2EDuration="3.772512099s" podCreationTimestamp="2025-12-03 18:45:53 +0000 UTC" firstStartedPulling="2025-12-03 18:45:54.968502488 +0000 UTC m=+6610.169879349" lastFinishedPulling="2025-12-03 18:45:55.549242906 +0000 UTC m=+6610.750619767" observedRunningTime="2025-12-03 18:45:56.762385645 +0000 UTC m=+6611.963762506" watchObservedRunningTime="2025-12-03 18:45:56.772512099 +0000 UTC m=+6611.973888960" Dec 03 18:45:56 crc kubenswrapper[4758]: I1203 18:45:56.786224 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" podStartSLOduration=3.786207858 podStartE2EDuration="3.786207858s" podCreationTimestamp="2025-12-03 18:45:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:45:56.779263 +0000 UTC m=+6611.980639861" watchObservedRunningTime="2025-12-03 18:45:56.786207858 +0000 UTC m=+6611.987584719" Dec 03 18:45:57 crc kubenswrapper[4758]: I1203 18:45:57.717342 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 18:45:58 crc kubenswrapper[4758]: I1203 18:45:58.114717 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:45:58 crc kubenswrapper[4758]: E1203 18:45:58.115363 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:46:03 crc kubenswrapper[4758]: I1203 18:46:03.872987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e7f8de8f-60ad-4b7e-92c5-6df43aae3589","Type":"ContainerStarted","Data":"f932317c88af15f108242fff8a9ca57809d3dad794530515e969b83b4e8dfea5"} Dec 03 18:46:03 crc kubenswrapper[4758]: I1203 18:46:03.873504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e7f8de8f-60ad-4b7e-92c5-6df43aae3589","Type":"ContainerStarted","Data":"b612d25372bbfa26aaa241171e99e1a22509cad32dccb828fbb940e5df339ed1"} Dec 03 18:46:03 crc kubenswrapper[4758]: I1203 18:46:03.905466 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.377550099 podStartE2EDuration="10.905447988s" podCreationTimestamp="2025-12-03 18:45:53 +0000 UTC" firstStartedPulling="2025-12-03 18:45:55.314805575 +0000 UTC m=+6610.516182436" lastFinishedPulling="2025-12-03 18:46:02.842703464 +0000 UTC m=+6618.044080325" observedRunningTime="2025-12-03 18:46:03.893815275 +0000 UTC m=+6619.095192156" watchObservedRunningTime="2025-12-03 18:46:03.905447988 +0000 UTC m=+6619.106824849" Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.174442 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.435218 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.460894 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.531160 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7446b96cf-q5w56"] Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.531399 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerName="dnsmasq-dns" containerID="cri-o://e61e072ca376bf4682ba7840dcbb62a3e7f56a0947d00c64d0333e16be43e17b" gracePeriod=10 Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.895271 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerID="e61e072ca376bf4682ba7840dcbb62a3e7f56a0947d00c64d0333e16be43e17b" exitCode=0 Dec 03 18:46:04 crc kubenswrapper[4758]: I1203 18:46:04.896586 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" event={"ID":"7a67effd-e5a5-4942-ae89-c717111fe61d","Type":"ContainerDied","Data":"e61e072ca376bf4682ba7840dcbb62a3e7f56a0947d00c64d0333e16be43e17b"} Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.102186 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.167087 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-dns-svc\") pod \"7a67effd-e5a5-4942-ae89-c717111fe61d\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.167237 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-config\") pod \"7a67effd-e5a5-4942-ae89-c717111fe61d\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.167295 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-nb\") pod \"7a67effd-e5a5-4942-ae89-c717111fe61d\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.167369 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlsgp\" (UniqueName: \"kubernetes.io/projected/7a67effd-e5a5-4942-ae89-c717111fe61d-kube-api-access-jlsgp\") pod \"7a67effd-e5a5-4942-ae89-c717111fe61d\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.167473 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-sb\") pod \"7a67effd-e5a5-4942-ae89-c717111fe61d\" (UID: \"7a67effd-e5a5-4942-ae89-c717111fe61d\") " Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.187591 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a67effd-e5a5-4942-ae89-c717111fe61d-kube-api-access-jlsgp" (OuterVolumeSpecName: "kube-api-access-jlsgp") pod "7a67effd-e5a5-4942-ae89-c717111fe61d" (UID: "7a67effd-e5a5-4942-ae89-c717111fe61d"). InnerVolumeSpecName "kube-api-access-jlsgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.247066 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a67effd-e5a5-4942-ae89-c717111fe61d" (UID: "7a67effd-e5a5-4942-ae89-c717111fe61d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.258910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a67effd-e5a5-4942-ae89-c717111fe61d" (UID: "7a67effd-e5a5-4942-ae89-c717111fe61d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.260234 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a67effd-e5a5-4942-ae89-c717111fe61d" (UID: "7a67effd-e5a5-4942-ae89-c717111fe61d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.269877 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.269907 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.269918 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.269927 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlsgp\" (UniqueName: \"kubernetes.io/projected/7a67effd-e5a5-4942-ae89-c717111fe61d-kube-api-access-jlsgp\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.276926 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-config" (OuterVolumeSpecName: "config") pod "7a67effd-e5a5-4942-ae89-c717111fe61d" (UID: "7a67effd-e5a5-4942-ae89-c717111fe61d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.371810 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a67effd-e5a5-4942-ae89-c717111fe61d-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.917326 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.918931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7446b96cf-q5w56" event={"ID":"7a67effd-e5a5-4942-ae89-c717111fe61d","Type":"ContainerDied","Data":"9827509f631626921a9ef3c03e6fe238e1b32e06f0d02a13403031c881f86b65"} Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.918991 4758 scope.go:117] "RemoveContainer" containerID="e61e072ca376bf4682ba7840dcbb62a3e7f56a0947d00c64d0333e16be43e17b" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.957010 4758 scope.go:117] "RemoveContainer" containerID="9c2d3cf728b689802439f5a17fc863ef657bf5d9af0d84183e8641fce0557714" Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.961883 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7446b96cf-q5w56"] Dec 03 18:46:05 crc kubenswrapper[4758]: I1203 18:46:05.971502 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7446b96cf-q5w56"] Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.127982 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" path="/var/lib/kubelet/pods/7a67effd-e5a5-4942-ae89-c717111fe61d/volumes" Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.419649 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.419930 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-central-agent" containerID="cri-o://bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331" gracePeriod=30 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.419997 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="proxy-httpd" containerID="cri-o://4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21" gracePeriod=30 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.420050 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-notification-agent" containerID="cri-o://cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24" gracePeriod=30 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.420058 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="sg-core" containerID="cri-o://ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27" gracePeriod=30 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.966892 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerID="4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21" exitCode=0 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.967232 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerID="ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27" exitCode=2 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.967243 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerID="bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331" exitCode=0 Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.967286 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerDied","Data":"4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21"} Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.967315 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerDied","Data":"ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27"} Dec 03 18:46:07 crc kubenswrapper[4758]: I1203 18:46:07.967331 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerDied","Data":"bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331"} Dec 03 18:46:11 crc kubenswrapper[4758]: I1203 18:46:11.115423 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:46:11 crc kubenswrapper[4758]: E1203 18:46:11.116350 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:46:13 crc kubenswrapper[4758]: I1203 18:46:13.931240 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.041367 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerID="cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24" exitCode=0 Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.041867 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerDied","Data":"cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24"} Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.041904 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a99a1df-a9fb-4906-bc9d-40a814eec360","Type":"ContainerDied","Data":"faa1ae522e67c0c0a079ac85d80d3d5e571f8999994e3ab50ea1b24e365cba06"} Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.041924 4758 scope.go:117] "RemoveContainer" containerID="4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.042088 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051470 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-config-data\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051528 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-combined-ca-bundle\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051574 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-sg-core-conf-yaml\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051616 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-log-httpd\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051863 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-run-httpd\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051898 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-scripts\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.051934 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd4r8\" (UniqueName: \"kubernetes.io/projected/8a99a1df-a9fb-4906-bc9d-40a814eec360-kube-api-access-vd4r8\") pod \"8a99a1df-a9fb-4906-bc9d-40a814eec360\" (UID: \"8a99a1df-a9fb-4906-bc9d-40a814eec360\") " Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.052332 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.052412 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.052456 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.071570 4758 scope.go:117] "RemoveContainer" containerID="ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.072634 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a99a1df-a9fb-4906-bc9d-40a814eec360-kube-api-access-vd4r8" (OuterVolumeSpecName: "kube-api-access-vd4r8") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "kube-api-access-vd4r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.084994 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-scripts" (OuterVolumeSpecName: "scripts") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.095962 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.157403 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.157442 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd4r8\" (UniqueName: \"kubernetes.io/projected/8a99a1df-a9fb-4906-bc9d-40a814eec360-kube-api-access-vd4r8\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.157456 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.157464 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a99a1df-a9fb-4906-bc9d-40a814eec360-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.170839 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.186850 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-config-data" (OuterVolumeSpecName: "config-data") pod "8a99a1df-a9fb-4906-bc9d-40a814eec360" (UID: "8a99a1df-a9fb-4906-bc9d-40a814eec360"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.204424 4758 scope.go:117] "RemoveContainer" containerID="cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.223030 4758 scope.go:117] "RemoveContainer" containerID="bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.244773 4758 scope.go:117] "RemoveContainer" containerID="4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.245123 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21\": container with ID starting with 4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21 not found: ID does not exist" containerID="4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245152 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21"} err="failed to get container status \"4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21\": rpc error: code = NotFound desc = could not find container \"4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21\": container with ID starting with 4a211e24a1d3b748f9988806718a83b7c6af39b3b486cb483127cf8fb12b5a21 not found: ID does not exist" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245173 4758 scope.go:117] "RemoveContainer" containerID="ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.245367 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27\": container with ID starting with ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27 not found: ID does not exist" containerID="ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245389 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27"} err="failed to get container status \"ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27\": rpc error: code = NotFound desc = could not find container \"ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27\": container with ID starting with ba49d2145f78d4277b640122d4516da3a155f2ff1e749685ca035ffaf4d35a27 not found: ID does not exist" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245401 4758 scope.go:117] "RemoveContainer" containerID="cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.245652 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24\": container with ID starting with cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24 not found: ID does not exist" containerID="cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245673 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24"} err="failed to get container status \"cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24\": rpc error: code = NotFound desc = could not find container \"cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24\": container with ID starting with cbf8b791f2fee52bf3fb8a09186171dec22e2b8fe24cf8bce794113c64a04b24 not found: ID does not exist" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245699 4758 scope.go:117] "RemoveContainer" containerID="bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.245876 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331\": container with ID starting with bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331 not found: ID does not exist" containerID="bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.245904 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331"} err="failed to get container status \"bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331\": rpc error: code = NotFound desc = could not find container \"bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331\": container with ID starting with bc5bbb14f5616c38c902e3e77b8a62f7d5b8e74dc0955fd99abb50c9a91cf331 not found: ID does not exist" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.259333 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.259364 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a99a1df-a9fb-4906-bc9d-40a814eec360-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.381131 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.391363 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.408687 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.409213 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerName="dnsmasq-dns" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409229 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerName="dnsmasq-dns" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.409255 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-central-agent" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409262 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-central-agent" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.409280 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="sg-core" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409286 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="sg-core" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.409305 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerName="init" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409310 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerName="init" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.409325 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="proxy-httpd" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409330 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="proxy-httpd" Dec 03 18:46:14 crc kubenswrapper[4758]: E1203 18:46:14.409338 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-notification-agent" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409346 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-notification-agent" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409670 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="proxy-httpd" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409735 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="sg-core" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409745 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a67effd-e5a5-4942-ae89-c717111fe61d" containerName="dnsmasq-dns" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409757 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-notification-agent" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.409791 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" containerName="ceilometer-central-agent" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.412028 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.415734 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.418319 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.424371 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.565435 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.565874 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-log-httpd\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.565942 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-config-data\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.566062 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq6l5\" (UniqueName: \"kubernetes.io/projected/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-kube-api-access-vq6l5\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.566213 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.566286 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-scripts\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.566397 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-run-httpd\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668410 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668796 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-log-httpd\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668824 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-config-data\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq6l5\" (UniqueName: \"kubernetes.io/projected/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-kube-api-access-vq6l5\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668909 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668939 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-scripts\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.668973 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-run-httpd\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.669384 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-log-httpd\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.670836 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-run-httpd\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.674391 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.674560 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.674633 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-scripts\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.674847 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-config-data\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.684181 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq6l5\" (UniqueName: \"kubernetes.io/projected/f81e7681-d858-4cf2-97dc-08a6e8c0b6f6-kube-api-access-vq6l5\") pod \"ceilometer-0\" (UID: \"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6\") " pod="openstack/ceilometer-0" Dec 03 18:46:14 crc kubenswrapper[4758]: I1203 18:46:14.736468 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:46:15 crc kubenswrapper[4758]: I1203 18:46:15.127128 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a99a1df-a9fb-4906-bc9d-40a814eec360" path="/var/lib/kubelet/pods/8a99a1df-a9fb-4906-bc9d-40a814eec360/volumes" Dec 03 18:46:15 crc kubenswrapper[4758]: I1203 18:46:15.243358 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:46:15 crc kubenswrapper[4758]: W1203 18:46:15.247003 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf81e7681_d858_4cf2_97dc_08a6e8c0b6f6.slice/crio-5fb15c961a9ed5c1559e0a5a1fe4c683f0e8d8dc1765a4e28b051c14ff1c3647 WatchSource:0}: Error finding container 5fb15c961a9ed5c1559e0a5a1fe4c683f0e8d8dc1765a4e28b051c14ff1c3647: Status 404 returned error can't find the container with id 5fb15c961a9ed5c1559e0a5a1fe4c683f0e8d8dc1765a4e28b051c14ff1c3647 Dec 03 18:46:16 crc kubenswrapper[4758]: I1203 18:46:16.012904 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 03 18:46:16 crc kubenswrapper[4758]: I1203 18:46:16.072217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerStarted","Data":"5fb15c961a9ed5c1559e0a5a1fe4c683f0e8d8dc1765a4e28b051c14ff1c3647"} Dec 03 18:46:16 crc kubenswrapper[4758]: I1203 18:46:16.114856 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 03 18:46:16 crc kubenswrapper[4758]: I1203 18:46:16.326914 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 03 18:46:17 crc kubenswrapper[4758]: I1203 18:46:17.082942 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerStarted","Data":"8ebfecd5373add76330f6d27551292cf180f8b1a1db95909f3583145b974a404"} Dec 03 18:46:17 crc kubenswrapper[4758]: I1203 18:46:17.083201 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerStarted","Data":"a643da5aad3d7b44feaa6b723d20522b0937a01b378d36bef4d3dc64f57b5896"} Dec 03 18:46:18 crc kubenswrapper[4758]: I1203 18:46:18.100246 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerStarted","Data":"1bc1444935a382f3ffc78892a0def2007f54bf87083763f081561a2ece7872f0"} Dec 03 18:46:20 crc kubenswrapper[4758]: I1203 18:46:20.119608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerStarted","Data":"c577eac27736d8507191e98d64f9932f8afc4bbbe36a8112ca9bf77e0620b120"} Dec 03 18:46:20 crc kubenswrapper[4758]: I1203 18:46:20.120343 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 18:46:20 crc kubenswrapper[4758]: I1203 18:46:20.146182 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.498448443 podStartE2EDuration="6.146164674s" podCreationTimestamp="2025-12-03 18:46:14 +0000 UTC" firstStartedPulling="2025-12-03 18:46:15.250035823 +0000 UTC m=+6630.451412684" lastFinishedPulling="2025-12-03 18:46:18.897752054 +0000 UTC m=+6634.099128915" observedRunningTime="2025-12-03 18:46:20.138406935 +0000 UTC m=+6635.339783806" watchObservedRunningTime="2025-12-03 18:46:20.146164674 +0000 UTC m=+6635.347541535" Dec 03 18:46:23 crc kubenswrapper[4758]: I1203 18:46:23.115345 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:46:23 crc kubenswrapper[4758]: E1203 18:46:23.116357 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:46:36 crc kubenswrapper[4758]: I1203 18:46:36.114544 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:46:36 crc kubenswrapper[4758]: E1203 18:46:36.115256 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:46:44 crc kubenswrapper[4758]: I1203 18:46:44.742290 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 18:46:47 crc kubenswrapper[4758]: I1203 18:46:47.115220 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:46:47 crc kubenswrapper[4758]: E1203 18:46:47.116514 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:47:00 crc kubenswrapper[4758]: I1203 18:47:00.114176 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:47:00 crc kubenswrapper[4758]: E1203 18:47:00.115206 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.494634 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-745d685f-kcvlr"] Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.498529 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.507407 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.516910 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-745d685f-kcvlr"] Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.586251 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlrf9\" (UniqueName: \"kubernetes.io/projected/b22e8487-78cd-494a-b981-dd6852e0e0b1-kube-api-access-hlrf9\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.586403 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-config\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.586469 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-nb\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.586539 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-sb\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.586634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-openstack-cell1\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.586874 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-dns-svc\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.688434 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlrf9\" (UniqueName: \"kubernetes.io/projected/b22e8487-78cd-494a-b981-dd6852e0e0b1-kube-api-access-hlrf9\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.688552 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-config\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.688609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-nb\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.688646 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-sb\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.688812 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-openstack-cell1\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.688863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-dns-svc\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.689771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-config\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.689871 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-sb\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.690161 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-openstack-cell1\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.690583 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-dns-svc\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.690626 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-nb\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.716575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlrf9\" (UniqueName: \"kubernetes.io/projected/b22e8487-78cd-494a-b981-dd6852e0e0b1-kube-api-access-hlrf9\") pod \"dnsmasq-dns-745d685f-kcvlr\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:05 crc kubenswrapper[4758]: I1203 18:47:05.825367 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:06 crc kubenswrapper[4758]: I1203 18:47:06.361367 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-745d685f-kcvlr"] Dec 03 18:47:06 crc kubenswrapper[4758]: I1203 18:47:06.592838 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745d685f-kcvlr" event={"ID":"b22e8487-78cd-494a-b981-dd6852e0e0b1","Type":"ContainerStarted","Data":"79371e206c1596441fe5c0ff1ff6c942850bdaa5edd35d2c9c48cc7ddba3c41a"} Dec 03 18:47:07 crc kubenswrapper[4758]: I1203 18:47:07.602847 4758 generic.go:334] "Generic (PLEG): container finished" podID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerID="a74e96854e3a712940410df7305551a73f697a73f5376200536c98b28a6597ab" exitCode=0 Dec 03 18:47:07 crc kubenswrapper[4758]: I1203 18:47:07.602937 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745d685f-kcvlr" event={"ID":"b22e8487-78cd-494a-b981-dd6852e0e0b1","Type":"ContainerDied","Data":"a74e96854e3a712940410df7305551a73f697a73f5376200536c98b28a6597ab"} Dec 03 18:47:08 crc kubenswrapper[4758]: I1203 18:47:08.615450 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745d685f-kcvlr" event={"ID":"b22e8487-78cd-494a-b981-dd6852e0e0b1","Type":"ContainerStarted","Data":"bd2a855e018c92f534f19a6df26cd536eab6df25b95d8bf01354dd5fc922a1e9"} Dec 03 18:47:08 crc kubenswrapper[4758]: I1203 18:47:08.615722 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:08 crc kubenswrapper[4758]: I1203 18:47:08.641132 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-745d685f-kcvlr" podStartSLOduration=3.641111147 podStartE2EDuration="3.641111147s" podCreationTimestamp="2025-12-03 18:47:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:47:08.634760216 +0000 UTC m=+6683.836137077" watchObservedRunningTime="2025-12-03 18:47:08.641111147 +0000 UTC m=+6683.842488008" Dec 03 18:47:14 crc kubenswrapper[4758]: I1203 18:47:14.114460 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:47:14 crc kubenswrapper[4758]: E1203 18:47:14.115347 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:47:15 crc kubenswrapper[4758]: I1203 18:47:15.827442 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:15 crc kubenswrapper[4758]: I1203 18:47:15.892828 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bdc5b97bf-pmfc8"] Dec 03 18:47:15 crc kubenswrapper[4758]: I1203 18:47:15.893054 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" podUID="5209a976-1aab-4069-b293-086b22734c5e" containerName="dnsmasq-dns" containerID="cri-o://cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256" gracePeriod=10 Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.074047 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8444f75b9f-zf8xz"] Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.076019 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.120956 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8444f75b9f-zf8xz"] Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.150536 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-ovsdbserver-nb\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.150743 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmmtv\" (UniqueName: \"kubernetes.io/projected/68c8bfdf-3a84-4c3b-8a8d-1df245208200-kube-api-access-vmmtv\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.150878 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-openstack-cell1\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.151082 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-ovsdbserver-sb\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.151348 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-dns-svc\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.151437 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-config\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.253093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-dns-svc\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.253965 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-dns-svc\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.254061 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-config\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.254241 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-ovsdbserver-nb\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.254559 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmmtv\" (UniqueName: \"kubernetes.io/projected/68c8bfdf-3a84-4c3b-8a8d-1df245208200-kube-api-access-vmmtv\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.254609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-openstack-cell1\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.254768 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-ovsdbserver-sb\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.254917 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-config\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.255088 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-ovsdbserver-nb\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.255327 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-openstack-cell1\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.255471 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68c8bfdf-3a84-4c3b-8a8d-1df245208200-ovsdbserver-sb\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.291580 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmmtv\" (UniqueName: \"kubernetes.io/projected/68c8bfdf-3a84-4c3b-8a8d-1df245208200-kube-api-access-vmmtv\") pod \"dnsmasq-dns-8444f75b9f-zf8xz\" (UID: \"68c8bfdf-3a84-4c3b-8a8d-1df245208200\") " pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.395329 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.541950 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.665756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-sb\") pod \"5209a976-1aab-4069-b293-086b22734c5e\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.665828 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-config\") pod \"5209a976-1aab-4069-b293-086b22734c5e\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.665875 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lfwp\" (UniqueName: \"kubernetes.io/projected/5209a976-1aab-4069-b293-086b22734c5e-kube-api-access-2lfwp\") pod \"5209a976-1aab-4069-b293-086b22734c5e\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.667068 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-dns-svc\") pod \"5209a976-1aab-4069-b293-086b22734c5e\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.667190 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-nb\") pod \"5209a976-1aab-4069-b293-086b22734c5e\" (UID: \"5209a976-1aab-4069-b293-086b22734c5e\") " Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.673281 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5209a976-1aab-4069-b293-086b22734c5e-kube-api-access-2lfwp" (OuterVolumeSpecName: "kube-api-access-2lfwp") pod "5209a976-1aab-4069-b293-086b22734c5e" (UID: "5209a976-1aab-4069-b293-086b22734c5e"). InnerVolumeSpecName "kube-api-access-2lfwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.707834 4758 generic.go:334] "Generic (PLEG): container finished" podID="5209a976-1aab-4069-b293-086b22734c5e" containerID="cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256" exitCode=0 Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.707876 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" event={"ID":"5209a976-1aab-4069-b293-086b22734c5e","Type":"ContainerDied","Data":"cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256"} Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.707903 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" event={"ID":"5209a976-1aab-4069-b293-086b22734c5e","Type":"ContainerDied","Data":"edd0ed161a35d66f0bcca28cc36beb93667ae7791259efeeb0834f34b93e2c5e"} Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.707918 4758 scope.go:117] "RemoveContainer" containerID="cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.708045 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdc5b97bf-pmfc8" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.725123 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-config" (OuterVolumeSpecName: "config") pod "5209a976-1aab-4069-b293-086b22734c5e" (UID: "5209a976-1aab-4069-b293-086b22734c5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.742267 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5209a976-1aab-4069-b293-086b22734c5e" (UID: "5209a976-1aab-4069-b293-086b22734c5e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.754070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5209a976-1aab-4069-b293-086b22734c5e" (UID: "5209a976-1aab-4069-b293-086b22734c5e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.756948 4758 scope.go:117] "RemoveContainer" containerID="4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.763504 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5209a976-1aab-4069-b293-086b22734c5e" (UID: "5209a976-1aab-4069-b293-086b22734c5e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.769852 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.769966 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.770047 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.770126 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lfwp\" (UniqueName: \"kubernetes.io/projected/5209a976-1aab-4069-b293-086b22734c5e-kube-api-access-2lfwp\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.770200 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5209a976-1aab-4069-b293-086b22734c5e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.777862 4758 scope.go:117] "RemoveContainer" containerID="cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256" Dec 03 18:47:16 crc kubenswrapper[4758]: E1203 18:47:16.778839 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256\": container with ID starting with cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256 not found: ID does not exist" containerID="cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.778886 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256"} err="failed to get container status \"cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256\": rpc error: code = NotFound desc = could not find container \"cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256\": container with ID starting with cbbdc09bd09aa84f9b973b4232facb6f88c27a91743e93a2a014c234006c5256 not found: ID does not exist" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.778914 4758 scope.go:117] "RemoveContainer" containerID="4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7" Dec 03 18:47:16 crc kubenswrapper[4758]: E1203 18:47:16.780088 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7\": container with ID starting with 4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7 not found: ID does not exist" containerID="4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.780116 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7"} err="failed to get container status \"4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7\": rpc error: code = NotFound desc = could not find container \"4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7\": container with ID starting with 4670d34c4f60693fdcd4eb02bf5c9f69deac0e15912eb6eaf85f04d700f80dc7 not found: ID does not exist" Dec 03 18:47:16 crc kubenswrapper[4758]: I1203 18:47:16.961972 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8444f75b9f-zf8xz"] Dec 03 18:47:17 crc kubenswrapper[4758]: I1203 18:47:17.185076 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bdc5b97bf-pmfc8"] Dec 03 18:47:17 crc kubenswrapper[4758]: I1203 18:47:17.194546 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bdc5b97bf-pmfc8"] Dec 03 18:47:17 crc kubenswrapper[4758]: I1203 18:47:17.720299 4758 generic.go:334] "Generic (PLEG): container finished" podID="68c8bfdf-3a84-4c3b-8a8d-1df245208200" containerID="926a278a2210f704a288ec0d8b440b52261e8c00f94610f6b62dd3f0fbc4f89a" exitCode=0 Dec 03 18:47:17 crc kubenswrapper[4758]: I1203 18:47:17.720354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" event={"ID":"68c8bfdf-3a84-4c3b-8a8d-1df245208200","Type":"ContainerDied","Data":"926a278a2210f704a288ec0d8b440b52261e8c00f94610f6b62dd3f0fbc4f89a"} Dec 03 18:47:17 crc kubenswrapper[4758]: I1203 18:47:17.720377 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" event={"ID":"68c8bfdf-3a84-4c3b-8a8d-1df245208200","Type":"ContainerStarted","Data":"a446df3d0f4cc77ee99f71e53d919364363bde3499f967178057437c7a72a2c5"} Dec 03 18:47:18 crc kubenswrapper[4758]: I1203 18:47:18.735267 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" event={"ID":"68c8bfdf-3a84-4c3b-8a8d-1df245208200","Type":"ContainerStarted","Data":"ea63a84fc424455d66f21e901187dd0683b13dfc537a68a5e11671c77195e123"} Dec 03 18:47:18 crc kubenswrapper[4758]: I1203 18:47:18.735532 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:18 crc kubenswrapper[4758]: I1203 18:47:18.764572 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" podStartSLOduration=2.764547227 podStartE2EDuration="2.764547227s" podCreationTimestamp="2025-12-03 18:47:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:47:18.757287312 +0000 UTC m=+6693.958664253" watchObservedRunningTime="2025-12-03 18:47:18.764547227 +0000 UTC m=+6693.965924108" Dec 03 18:47:19 crc kubenswrapper[4758]: I1203 18:47:19.124891 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5209a976-1aab-4069-b293-086b22734c5e" path="/var/lib/kubelet/pods/5209a976-1aab-4069-b293-086b22734c5e/volumes" Dec 03 18:47:26 crc kubenswrapper[4758]: I1203 18:47:26.397997 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8444f75b9f-zf8xz" Dec 03 18:47:26 crc kubenswrapper[4758]: I1203 18:47:26.561338 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-745d685f-kcvlr"] Dec 03 18:47:26 crc kubenswrapper[4758]: I1203 18:47:26.561578 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-745d685f-kcvlr" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerName="dnsmasq-dns" containerID="cri-o://bd2a855e018c92f534f19a6df26cd536eab6df25b95d8bf01354dd5fc922a1e9" gracePeriod=10 Dec 03 18:47:26 crc kubenswrapper[4758]: I1203 18:47:26.833915 4758 generic.go:334] "Generic (PLEG): container finished" podID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerID="bd2a855e018c92f534f19a6df26cd536eab6df25b95d8bf01354dd5fc922a1e9" exitCode=0 Dec 03 18:47:26 crc kubenswrapper[4758]: I1203 18:47:26.834156 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745d685f-kcvlr" event={"ID":"b22e8487-78cd-494a-b981-dd6852e0e0b1","Type":"ContainerDied","Data":"bd2a855e018c92f534f19a6df26cd536eab6df25b95d8bf01354dd5fc922a1e9"} Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.131278 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.215972 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlrf9\" (UniqueName: \"kubernetes.io/projected/b22e8487-78cd-494a-b981-dd6852e0e0b1-kube-api-access-hlrf9\") pod \"b22e8487-78cd-494a-b981-dd6852e0e0b1\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.216088 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-nb\") pod \"b22e8487-78cd-494a-b981-dd6852e0e0b1\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.216196 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-config\") pod \"b22e8487-78cd-494a-b981-dd6852e0e0b1\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.216979 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-openstack-cell1\") pod \"b22e8487-78cd-494a-b981-dd6852e0e0b1\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.217030 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-dns-svc\") pod \"b22e8487-78cd-494a-b981-dd6852e0e0b1\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.217228 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-sb\") pod \"b22e8487-78cd-494a-b981-dd6852e0e0b1\" (UID: \"b22e8487-78cd-494a-b981-dd6852e0e0b1\") " Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.238048 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b22e8487-78cd-494a-b981-dd6852e0e0b1-kube-api-access-hlrf9" (OuterVolumeSpecName: "kube-api-access-hlrf9") pod "b22e8487-78cd-494a-b981-dd6852e0e0b1" (UID: "b22e8487-78cd-494a-b981-dd6852e0e0b1"). InnerVolumeSpecName "kube-api-access-hlrf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.271367 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b22e8487-78cd-494a-b981-dd6852e0e0b1" (UID: "b22e8487-78cd-494a-b981-dd6852e0e0b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.275455 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b22e8487-78cd-494a-b981-dd6852e0e0b1" (UID: "b22e8487-78cd-494a-b981-dd6852e0e0b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.300242 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-config" (OuterVolumeSpecName: "config") pod "b22e8487-78cd-494a-b981-dd6852e0e0b1" (UID: "b22e8487-78cd-494a-b981-dd6852e0e0b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.300883 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b22e8487-78cd-494a-b981-dd6852e0e0b1" (UID: "b22e8487-78cd-494a-b981-dd6852e0e0b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.314860 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "b22e8487-78cd-494a-b981-dd6852e0e0b1" (UID: "b22e8487-78cd-494a-b981-dd6852e0e0b1"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.319888 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlrf9\" (UniqueName: \"kubernetes.io/projected/b22e8487-78cd-494a-b981-dd6852e0e0b1-kube-api-access-hlrf9\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.319921 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.319932 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.319944 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.319956 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.319963 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b22e8487-78cd-494a-b981-dd6852e0e0b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.848179 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745d685f-kcvlr" event={"ID":"b22e8487-78cd-494a-b981-dd6852e0e0b1","Type":"ContainerDied","Data":"79371e206c1596441fe5c0ff1ff6c942850bdaa5edd35d2c9c48cc7ddba3c41a"} Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.848481 4758 scope.go:117] "RemoveContainer" containerID="bd2a855e018c92f534f19a6df26cd536eab6df25b95d8bf01354dd5fc922a1e9" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.848332 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745d685f-kcvlr" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.881046 4758 scope.go:117] "RemoveContainer" containerID="a74e96854e3a712940410df7305551a73f697a73f5376200536c98b28a6597ab" Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.897703 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-745d685f-kcvlr"] Dec 03 18:47:27 crc kubenswrapper[4758]: I1203 18:47:27.912786 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-745d685f-kcvlr"] Dec 03 18:47:28 crc kubenswrapper[4758]: I1203 18:47:28.114990 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:47:28 crc kubenswrapper[4758]: E1203 18:47:28.115513 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:47:29 crc kubenswrapper[4758]: I1203 18:47:29.129255 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" path="/var/lib/kubelet/pods/b22e8487-78cd-494a-b981-dd6852e0e0b1/volumes" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.446738 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks"] Dec 03 18:47:37 crc kubenswrapper[4758]: E1203 18:47:37.464101 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerName="init" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.464166 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerName="init" Dec 03 18:47:37 crc kubenswrapper[4758]: E1203 18:47:37.464193 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5209a976-1aab-4069-b293-086b22734c5e" containerName="init" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.464199 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5209a976-1aab-4069-b293-086b22734c5e" containerName="init" Dec 03 18:47:37 crc kubenswrapper[4758]: E1203 18:47:37.464231 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5209a976-1aab-4069-b293-086b22734c5e" containerName="dnsmasq-dns" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.464237 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5209a976-1aab-4069-b293-086b22734c5e" containerName="dnsmasq-dns" Dec 03 18:47:37 crc kubenswrapper[4758]: E1203 18:47:37.464251 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerName="dnsmasq-dns" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.464257 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerName="dnsmasq-dns" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.464463 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5209a976-1aab-4069-b293-086b22734c5e" containerName="dnsmasq-dns" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.464473 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b22e8487-78cd-494a-b981-dd6852e0e0b1" containerName="dnsmasq-dns" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.465176 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks"] Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.465258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.469207 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.469460 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.470315 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.471461 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.559289 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fwtv\" (UniqueName: \"kubernetes.io/projected/92bcc00c-1a64-4bb1-8152-25f30ff242a9-kube-api-access-6fwtv\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.559899 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.559976 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.560033 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.560075 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.662453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.662607 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fwtv\" (UniqueName: \"kubernetes.io/projected/92bcc00c-1a64-4bb1-8152-25f30ff242a9-kube-api-access-6fwtv\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.662835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.662877 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.662912 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.669592 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.669626 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.670446 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.674203 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.679417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fwtv\" (UniqueName: \"kubernetes.io/projected/92bcc00c-1a64-4bb1-8152-25f30ff242a9-kube-api-access-6fwtv\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:37 crc kubenswrapper[4758]: I1203 18:47:37.801638 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:47:38 crc kubenswrapper[4758]: I1203 18:47:38.349707 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks"] Dec 03 18:47:38 crc kubenswrapper[4758]: I1203 18:47:38.972016 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" event={"ID":"92bcc00c-1a64-4bb1-8152-25f30ff242a9","Type":"ContainerStarted","Data":"6917d206fe4383441d8be8c585e1b84a351690ad51a923c5598590387c4ebfb8"} Dec 03 18:47:40 crc kubenswrapper[4758]: I1203 18:47:40.115298 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:47:40 crc kubenswrapper[4758]: E1203 18:47:40.116058 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:47:53 crc kubenswrapper[4758]: I1203 18:47:53.117723 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:47:53 crc kubenswrapper[4758]: E1203 18:47:53.118571 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:47:55 crc kubenswrapper[4758]: I1203 18:47:55.547491 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:47:56 crc kubenswrapper[4758]: I1203 18:47:56.053168 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-lgqzq"] Dec 03 18:47:56 crc kubenswrapper[4758]: I1203 18:47:56.064333 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-lgqzq"] Dec 03 18:47:56 crc kubenswrapper[4758]: I1203 18:47:56.139271 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" event={"ID":"92bcc00c-1a64-4bb1-8152-25f30ff242a9","Type":"ContainerStarted","Data":"21d43631eab8a6d9b17361df07ce0f59842b394582b5eb1f950e770ecbe942a9"} Dec 03 18:47:56 crc kubenswrapper[4758]: I1203 18:47:56.157153 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" podStartSLOduration=1.975307145 podStartE2EDuration="19.157134528s" podCreationTimestamp="2025-12-03 18:47:37 +0000 UTC" firstStartedPulling="2025-12-03 18:47:38.358117884 +0000 UTC m=+6713.559494745" lastFinishedPulling="2025-12-03 18:47:55.539945267 +0000 UTC m=+6730.741322128" observedRunningTime="2025-12-03 18:47:56.154429115 +0000 UTC m=+6731.355805996" watchObservedRunningTime="2025-12-03 18:47:56.157134528 +0000 UTC m=+6731.358511379" Dec 03 18:47:57 crc kubenswrapper[4758]: I1203 18:47:57.030913 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-b699-account-create-update-j7gtf"] Dec 03 18:47:57 crc kubenswrapper[4758]: I1203 18:47:57.041601 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-b699-account-create-update-j7gtf"] Dec 03 18:47:57 crc kubenswrapper[4758]: I1203 18:47:57.127292 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412a365c-25b6-4058-9b1f-be4f76154b6a" path="/var/lib/kubelet/pods/412a365c-25b6-4058-9b1f-be4f76154b6a/volumes" Dec 03 18:47:57 crc kubenswrapper[4758]: I1203 18:47:57.128481 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a25bce28-2142-4800-b3d6-e0059dae2e37" path="/var/lib/kubelet/pods/a25bce28-2142-4800-b3d6-e0059dae2e37/volumes" Dec 03 18:48:04 crc kubenswrapper[4758]: I1203 18:48:04.035610 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-w62m7"] Dec 03 18:48:04 crc kubenswrapper[4758]: I1203 18:48:04.047277 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-a842-account-create-update-h2jjh"] Dec 03 18:48:04 crc kubenswrapper[4758]: I1203 18:48:04.057200 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-w62m7"] Dec 03 18:48:04 crc kubenswrapper[4758]: I1203 18:48:04.066911 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-a842-account-create-update-h2jjh"] Dec 03 18:48:05 crc kubenswrapper[4758]: I1203 18:48:05.132272 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c1fd37-adec-4fa9-b28c-9258306a3ad0" path="/var/lib/kubelet/pods/87c1fd37-adec-4fa9-b28c-9258306a3ad0/volumes" Dec 03 18:48:05 crc kubenswrapper[4758]: I1203 18:48:05.133463 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a288340d-898d-414d-beea-b40379693a89" path="/var/lib/kubelet/pods/a288340d-898d-414d-beea-b40379693a89/volumes" Dec 03 18:48:06 crc kubenswrapper[4758]: I1203 18:48:06.114483 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:48:06 crc kubenswrapper[4758]: E1203 18:48:06.114773 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:48:09 crc kubenswrapper[4758]: I1203 18:48:09.260491 4758 generic.go:334] "Generic (PLEG): container finished" podID="92bcc00c-1a64-4bb1-8152-25f30ff242a9" containerID="21d43631eab8a6d9b17361df07ce0f59842b394582b5eb1f950e770ecbe942a9" exitCode=0 Dec 03 18:48:09 crc kubenswrapper[4758]: I1203 18:48:09.260573 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" event={"ID":"92bcc00c-1a64-4bb1-8152-25f30ff242a9","Type":"ContainerDied","Data":"21d43631eab8a6d9b17361df07ce0f59842b394582b5eb1f950e770ecbe942a9"} Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.782580 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.909200 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-inventory\") pod \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.909295 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-pre-adoption-validation-combined-ca-bundle\") pod \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.909342 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fwtv\" (UniqueName: \"kubernetes.io/projected/92bcc00c-1a64-4bb1-8152-25f30ff242a9-kube-api-access-6fwtv\") pod \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.909367 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ceph\") pod \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.909414 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ssh-key\") pod \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\" (UID: \"92bcc00c-1a64-4bb1-8152-25f30ff242a9\") " Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.916227 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ceph" (OuterVolumeSpecName: "ceph") pod "92bcc00c-1a64-4bb1-8152-25f30ff242a9" (UID: "92bcc00c-1a64-4bb1-8152-25f30ff242a9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.922744 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92bcc00c-1a64-4bb1-8152-25f30ff242a9-kube-api-access-6fwtv" (OuterVolumeSpecName: "kube-api-access-6fwtv") pod "92bcc00c-1a64-4bb1-8152-25f30ff242a9" (UID: "92bcc00c-1a64-4bb1-8152-25f30ff242a9"). InnerVolumeSpecName "kube-api-access-6fwtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.922896 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "92bcc00c-1a64-4bb1-8152-25f30ff242a9" (UID: "92bcc00c-1a64-4bb1-8152-25f30ff242a9"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.943350 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-inventory" (OuterVolumeSpecName: "inventory") pod "92bcc00c-1a64-4bb1-8152-25f30ff242a9" (UID: "92bcc00c-1a64-4bb1-8152-25f30ff242a9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:48:10 crc kubenswrapper[4758]: I1203 18:48:10.948372 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92bcc00c-1a64-4bb1-8152-25f30ff242a9" (UID: "92bcc00c-1a64-4bb1-8152-25f30ff242a9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.012757 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.012807 4758 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.012823 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fwtv\" (UniqueName: \"kubernetes.io/projected/92bcc00c-1a64-4bb1-8152-25f30ff242a9-kube-api-access-6fwtv\") on node \"crc\" DevicePath \"\"" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.012834 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.012842 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92bcc00c-1a64-4bb1-8152-25f30ff242a9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.289588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" event={"ID":"92bcc00c-1a64-4bb1-8152-25f30ff242a9","Type":"ContainerDied","Data":"6917d206fe4383441d8be8c585e1b84a351690ad51a923c5598590387c4ebfb8"} Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.289658 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6917d206fe4383441d8be8c585e1b84a351690ad51a923c5598590387c4ebfb8" Dec 03 18:48:11 crc kubenswrapper[4758]: I1203 18:48:11.289669 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks" Dec 03 18:48:15 crc kubenswrapper[4758]: I1203 18:48:15.550467 4758 scope.go:117] "RemoveContainer" containerID="7412f585940e1590344ccaa4e9a9d6f736a540717529cbaf5945b46388574fbb" Dec 03 18:48:15 crc kubenswrapper[4758]: I1203 18:48:15.574873 4758 scope.go:117] "RemoveContainer" containerID="c337c8fd45c5133eac7b793fb1afc59822251f795242a8372cacdebcba091999" Dec 03 18:48:15 crc kubenswrapper[4758]: I1203 18:48:15.612450 4758 scope.go:117] "RemoveContainer" containerID="2850d6ac7f4fcf029b36d31c28e133c3c6e1c57a50afaf1e0b39293f115e1c33" Dec 03 18:48:15 crc kubenswrapper[4758]: I1203 18:48:15.821911 4758 scope.go:117] "RemoveContainer" containerID="7b5d0b89ed397f2b8c6f3b5e16fc897405d08d2a66d33867d44c7efb9e88722f" Dec 03 18:48:15 crc kubenswrapper[4758]: I1203 18:48:15.862144 4758 scope.go:117] "RemoveContainer" containerID="7bc767c44181b8918de1c95da92754abaacf5735a48379973f49890aec4ab3d2" Dec 03 18:48:15 crc kubenswrapper[4758]: I1203 18:48:15.913954 4758 scope.go:117] "RemoveContainer" containerID="e4d51b7c62d37912e8953538b9be10602062551fee21f16298b3a19549882db7" Dec 03 18:48:17 crc kubenswrapper[4758]: I1203 18:48:17.115626 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:48:17 crc kubenswrapper[4758]: E1203 18:48:17.116282 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.255244 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m"] Dec 03 18:48:21 crc kubenswrapper[4758]: E1203 18:48:21.256365 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bcc00c-1a64-4bb1-8152-25f30ff242a9" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.256385 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bcc00c-1a64-4bb1-8152-25f30ff242a9" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.256653 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="92bcc00c-1a64-4bb1-8152-25f30ff242a9" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.258953 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.261971 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.262273 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.262419 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.264094 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.267544 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m"] Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.370831 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.371284 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.371444 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.371574 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxxft\" (UniqueName: \"kubernetes.io/projected/650575e1-2ffe-41e4-887a-ae4e451bc53a-kube-api-access-qxxft\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.371722 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.473848 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.473910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxxft\" (UniqueName: \"kubernetes.io/projected/650575e1-2ffe-41e4-887a-ae4e451bc53a-kube-api-access-qxxft\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.473957 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.474088 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.474114 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.484302 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.484411 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.484579 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.484639 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.490581 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxxft\" (UniqueName: \"kubernetes.io/projected/650575e1-2ffe-41e4-887a-ae4e451bc53a-kube-api-access-qxxft\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:21 crc kubenswrapper[4758]: I1203 18:48:21.586503 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:48:22 crc kubenswrapper[4758]: I1203 18:48:22.263087 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m"] Dec 03 18:48:22 crc kubenswrapper[4758]: W1203 18:48:22.263726 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod650575e1_2ffe_41e4_887a_ae4e451bc53a.slice/crio-b7fcf5e6edc1dcef8477d35ff29bc61d310c13a0674fd405f0718b02346b54df WatchSource:0}: Error finding container b7fcf5e6edc1dcef8477d35ff29bc61d310c13a0674fd405f0718b02346b54df: Status 404 returned error can't find the container with id b7fcf5e6edc1dcef8477d35ff29bc61d310c13a0674fd405f0718b02346b54df Dec 03 18:48:22 crc kubenswrapper[4758]: I1203 18:48:22.413893 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" event={"ID":"650575e1-2ffe-41e4-887a-ae4e451bc53a","Type":"ContainerStarted","Data":"b7fcf5e6edc1dcef8477d35ff29bc61d310c13a0674fd405f0718b02346b54df"} Dec 03 18:48:23 crc kubenswrapper[4758]: I1203 18:48:23.426078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" event={"ID":"650575e1-2ffe-41e4-887a-ae4e451bc53a","Type":"ContainerStarted","Data":"c1abc6ce129c0e89d9c1a1c461158862c0cc032849fb1867590ee158d2adf075"} Dec 03 18:48:23 crc kubenswrapper[4758]: I1203 18:48:23.444111 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" podStartSLOduration=2.121513738 podStartE2EDuration="2.444094795s" podCreationTimestamp="2025-12-03 18:48:21 +0000 UTC" firstStartedPulling="2025-12-03 18:48:22.267556214 +0000 UTC m=+6757.468933075" lastFinishedPulling="2025-12-03 18:48:22.590137271 +0000 UTC m=+6757.791514132" observedRunningTime="2025-12-03 18:48:23.441543686 +0000 UTC m=+6758.642920557" watchObservedRunningTime="2025-12-03 18:48:23.444094795 +0000 UTC m=+6758.645471656" Dec 03 18:48:30 crc kubenswrapper[4758]: I1203 18:48:30.114112 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:48:30 crc kubenswrapper[4758]: E1203 18:48:30.114790 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:48:44 crc kubenswrapper[4758]: I1203 18:48:44.114482 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:48:44 crc kubenswrapper[4758]: E1203 18:48:44.115318 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:48:58 crc kubenswrapper[4758]: I1203 18:48:58.115338 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:48:58 crc kubenswrapper[4758]: E1203 18:48:58.116297 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:49:12 crc kubenswrapper[4758]: I1203 18:49:12.115060 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:49:12 crc kubenswrapper[4758]: I1203 18:49:12.949567 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"64411234ffdaf6687b4f6e58c921ca9496d13f2f2e70093e2409ece7a93f666b"} Dec 03 18:49:21 crc kubenswrapper[4758]: I1203 18:49:21.042732 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-gpjpc"] Dec 03 18:49:21 crc kubenswrapper[4758]: I1203 18:49:21.051973 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-gpjpc"] Dec 03 18:49:21 crc kubenswrapper[4758]: I1203 18:49:21.128019 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea3c817-521f-4709-b2c2-ab41fe7d9c81" path="/var/lib/kubelet/pods/dea3c817-521f-4709-b2c2-ab41fe7d9c81/volumes" Dec 03 18:50:16 crc kubenswrapper[4758]: I1203 18:50:16.094784 4758 scope.go:117] "RemoveContainer" containerID="d583bce5e20757c772d5d97f6b801869050141e60bbde59f7324fe621b40ecba" Dec 03 18:50:16 crc kubenswrapper[4758]: I1203 18:50:16.121147 4758 scope.go:117] "RemoveContainer" containerID="b281515c0cd165d6b3a31178a814fabae1924cebc0c8c28f3f840224fb5de799" Dec 03 18:51:41 crc kubenswrapper[4758]: I1203 18:51:41.394449 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:51:41 crc kubenswrapper[4758]: I1203 18:51:41.395069 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:52:11 crc kubenswrapper[4758]: I1203 18:52:11.395246 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:52:11 crc kubenswrapper[4758]: I1203 18:52:11.395869 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:52:36 crc kubenswrapper[4758]: I1203 18:52:36.041955 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-5c0a-account-create-update-dmf7v"] Dec 03 18:52:36 crc kubenswrapper[4758]: I1203 18:52:36.051236 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-6kbkf"] Dec 03 18:52:36 crc kubenswrapper[4758]: I1203 18:52:36.061017 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-6kbkf"] Dec 03 18:52:36 crc kubenswrapper[4758]: I1203 18:52:36.069268 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-5c0a-account-create-update-dmf7v"] Dec 03 18:52:37 crc kubenswrapper[4758]: I1203 18:52:37.127532 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5" path="/var/lib/kubelet/pods/6c2271b8-acf4-4049-a1cd-93a4a7b5b2c5/volumes" Dec 03 18:52:37 crc kubenswrapper[4758]: I1203 18:52:37.129011 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e4a382-3f17-482b-a2c2-cbdc11dbef9c" path="/var/lib/kubelet/pods/b3e4a382-3f17-482b-a2c2-cbdc11dbef9c/volumes" Dec 03 18:52:41 crc kubenswrapper[4758]: I1203 18:52:41.395203 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:52:41 crc kubenswrapper[4758]: I1203 18:52:41.396137 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:52:41 crc kubenswrapper[4758]: I1203 18:52:41.397118 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:52:41 crc kubenswrapper[4758]: I1203 18:52:41.398743 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64411234ffdaf6687b4f6e58c921ca9496d13f2f2e70093e2409ece7a93f666b"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:52:41 crc kubenswrapper[4758]: I1203 18:52:41.398869 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://64411234ffdaf6687b4f6e58c921ca9496d13f2f2e70093e2409ece7a93f666b" gracePeriod=600 Dec 03 18:52:42 crc kubenswrapper[4758]: I1203 18:52:42.508231 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="64411234ffdaf6687b4f6e58c921ca9496d13f2f2e70093e2409ece7a93f666b" exitCode=0 Dec 03 18:52:42 crc kubenswrapper[4758]: I1203 18:52:42.508319 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"64411234ffdaf6687b4f6e58c921ca9496d13f2f2e70093e2409ece7a93f666b"} Dec 03 18:52:42 crc kubenswrapper[4758]: I1203 18:52:42.508834 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74"} Dec 03 18:52:42 crc kubenswrapper[4758]: I1203 18:52:42.508856 4758 scope.go:117] "RemoveContainer" containerID="59f604c4c241ff14dd841d3d31b9856d247fc5b7270e98a0f31e2d89ba380b00" Dec 03 18:52:50 crc kubenswrapper[4758]: I1203 18:52:50.057525 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-k7jtl"] Dec 03 18:52:50 crc kubenswrapper[4758]: I1203 18:52:50.069133 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-k7jtl"] Dec 03 18:52:51 crc kubenswrapper[4758]: I1203 18:52:51.135898 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2205b3e1-3e92-4688-a666-a730dd1cce69" path="/var/lib/kubelet/pods/2205b3e1-3e92-4688-a666-a730dd1cce69/volumes" Dec 03 18:53:15 crc kubenswrapper[4758]: I1203 18:53:15.892199 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ts8fz"] Dec 03 18:53:15 crc kubenswrapper[4758]: I1203 18:53:15.894970 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:15 crc kubenswrapper[4758]: I1203 18:53:15.902042 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ts8fz"] Dec 03 18:53:15 crc kubenswrapper[4758]: I1203 18:53:15.985623 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-utilities\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:15 crc kubenswrapper[4758]: I1203 18:53:15.986169 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d99xp\" (UniqueName: \"kubernetes.io/projected/2fab9f64-5948-43bd-8af4-4c3c60781792-kube-api-access-d99xp\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:15 crc kubenswrapper[4758]: I1203 18:53:15.986255 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-catalog-content\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.088450 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-utilities\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.088585 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d99xp\" (UniqueName: \"kubernetes.io/projected/2fab9f64-5948-43bd-8af4-4c3c60781792-kube-api-access-d99xp\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.088639 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-catalog-content\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.089203 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-utilities\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.089212 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-catalog-content\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.121672 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d99xp\" (UniqueName: \"kubernetes.io/projected/2fab9f64-5948-43bd-8af4-4c3c60781792-kube-api-access-d99xp\") pod \"community-operators-ts8fz\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.266512 4758 scope.go:117] "RemoveContainer" containerID="381a57e3ee67f4f00bb58b989420683a12c1a27a5fd06cec05ba15186632a020" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.274834 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.291602 4758 scope.go:117] "RemoveContainer" containerID="b39142ea20bffd6d24ece9d10ad0e515a58f35e65400ed8aad77a8ab4c287405" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.346921 4758 scope.go:117] "RemoveContainer" containerID="4497e840ec067ef7d5272d120eeb4b11167c1074923c2dc3726155ac792c5ef2" Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.863453 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ts8fz"] Dec 03 18:53:16 crc kubenswrapper[4758]: W1203 18:53:16.871912 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fab9f64_5948_43bd_8af4_4c3c60781792.slice/crio-1ad418e10699261d124c28995bafd807e6cd3be269aa68673e5a7778295cb311 WatchSource:0}: Error finding container 1ad418e10699261d124c28995bafd807e6cd3be269aa68673e5a7778295cb311: Status 404 returned error can't find the container with id 1ad418e10699261d124c28995bafd807e6cd3be269aa68673e5a7778295cb311 Dec 03 18:53:16 crc kubenswrapper[4758]: I1203 18:53:16.886896 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerStarted","Data":"1ad418e10699261d124c28995bafd807e6cd3be269aa68673e5a7778295cb311"} Dec 03 18:53:17 crc kubenswrapper[4758]: I1203 18:53:17.900385 4758 generic.go:334] "Generic (PLEG): container finished" podID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerID="93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d" exitCode=0 Dec 03 18:53:17 crc kubenswrapper[4758]: I1203 18:53:17.900443 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerDied","Data":"93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d"} Dec 03 18:53:17 crc kubenswrapper[4758]: I1203 18:53:17.902995 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.892384 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x44mg"] Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.895494 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.922512 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x44mg"] Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.931813 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerStarted","Data":"129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060"} Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.955224 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-utilities\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.955451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66xjt\" (UniqueName: \"kubernetes.io/projected/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-kube-api-access-66xjt\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:18 crc kubenswrapper[4758]: I1203 18:53:18.955485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-catalog-content\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.057503 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66xjt\" (UniqueName: \"kubernetes.io/projected/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-kube-api-access-66xjt\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.057552 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-catalog-content\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.057705 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-utilities\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.058231 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-catalog-content\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.058317 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-utilities\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.085371 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66xjt\" (UniqueName: \"kubernetes.io/projected/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-kube-api-access-66xjt\") pod \"redhat-operators-x44mg\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.229590 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.692463 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x44mg"] Dec 03 18:53:19 crc kubenswrapper[4758]: W1203 18:53:19.740894 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58c93d36_9a52_4a2e_96b4_d3bfd2a8d972.slice/crio-2447923717cbcbd40abaf6e1083399eb79c3a549c84979bb04bd01104d556e3a WatchSource:0}: Error finding container 2447923717cbcbd40abaf6e1083399eb79c3a549c84979bb04bd01104d556e3a: Status 404 returned error can't find the container with id 2447923717cbcbd40abaf6e1083399eb79c3a549c84979bb04bd01104d556e3a Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.944519 4758 generic.go:334] "Generic (PLEG): container finished" podID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerID="129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060" exitCode=0 Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.944574 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerDied","Data":"129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060"} Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.948338 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerStarted","Data":"c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83"} Dec 03 18:53:19 crc kubenswrapper[4758]: I1203 18:53:19.948371 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerStarted","Data":"2447923717cbcbd40abaf6e1083399eb79c3a549c84979bb04bd01104d556e3a"} Dec 03 18:53:20 crc kubenswrapper[4758]: I1203 18:53:20.958460 4758 generic.go:334] "Generic (PLEG): container finished" podID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerID="c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83" exitCode=0 Dec 03 18:53:20 crc kubenswrapper[4758]: I1203 18:53:20.958594 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerDied","Data":"c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83"} Dec 03 18:53:20 crc kubenswrapper[4758]: I1203 18:53:20.963291 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerStarted","Data":"461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe"} Dec 03 18:53:21 crc kubenswrapper[4758]: I1203 18:53:21.002990 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ts8fz" podStartSLOduration=3.567864245 podStartE2EDuration="6.002972836s" podCreationTimestamp="2025-12-03 18:53:15 +0000 UTC" firstStartedPulling="2025-12-03 18:53:17.902659742 +0000 UTC m=+7053.104036613" lastFinishedPulling="2025-12-03 18:53:20.337768343 +0000 UTC m=+7055.539145204" observedRunningTime="2025-12-03 18:53:20.998737662 +0000 UTC m=+7056.200114543" watchObservedRunningTime="2025-12-03 18:53:21.002972836 +0000 UTC m=+7056.204349697" Dec 03 18:53:21 crc kubenswrapper[4758]: I1203 18:53:21.979238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerStarted","Data":"cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca"} Dec 03 18:53:26 crc kubenswrapper[4758]: I1203 18:53:26.275395 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:26 crc kubenswrapper[4758]: I1203 18:53:26.276043 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:26 crc kubenswrapper[4758]: I1203 18:53:26.353066 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:27 crc kubenswrapper[4758]: I1203 18:53:27.073355 4758 generic.go:334] "Generic (PLEG): container finished" podID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerID="cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca" exitCode=0 Dec 03 18:53:27 crc kubenswrapper[4758]: I1203 18:53:27.073475 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerDied","Data":"cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca"} Dec 03 18:53:27 crc kubenswrapper[4758]: I1203 18:53:27.147005 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:28 crc kubenswrapper[4758]: I1203 18:53:28.074225 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ts8fz"] Dec 03 18:53:28 crc kubenswrapper[4758]: I1203 18:53:28.089496 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerStarted","Data":"350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5"} Dec 03 18:53:28 crc kubenswrapper[4758]: I1203 18:53:28.115298 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x44mg" podStartSLOduration=3.271430724 podStartE2EDuration="10.115279812s" podCreationTimestamp="2025-12-03 18:53:18 +0000 UTC" firstStartedPulling="2025-12-03 18:53:20.960513309 +0000 UTC m=+7056.161890180" lastFinishedPulling="2025-12-03 18:53:27.804362367 +0000 UTC m=+7063.005739268" observedRunningTime="2025-12-03 18:53:28.105800446 +0000 UTC m=+7063.307177327" watchObservedRunningTime="2025-12-03 18:53:28.115279812 +0000 UTC m=+7063.316656673" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.101206 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ts8fz" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="registry-server" containerID="cri-o://461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe" gracePeriod=2 Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.230958 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.231017 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.790973 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.904288 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d99xp\" (UniqueName: \"kubernetes.io/projected/2fab9f64-5948-43bd-8af4-4c3c60781792-kube-api-access-d99xp\") pod \"2fab9f64-5948-43bd-8af4-4c3c60781792\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.904437 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-catalog-content\") pod \"2fab9f64-5948-43bd-8af4-4c3c60781792\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.904530 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-utilities\") pod \"2fab9f64-5948-43bd-8af4-4c3c60781792\" (UID: \"2fab9f64-5948-43bd-8af4-4c3c60781792\") " Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.904977 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-utilities" (OuterVolumeSpecName: "utilities") pod "2fab9f64-5948-43bd-8af4-4c3c60781792" (UID: "2fab9f64-5948-43bd-8af4-4c3c60781792"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.905235 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.912154 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fab9f64-5948-43bd-8af4-4c3c60781792-kube-api-access-d99xp" (OuterVolumeSpecName: "kube-api-access-d99xp") pod "2fab9f64-5948-43bd-8af4-4c3c60781792" (UID: "2fab9f64-5948-43bd-8af4-4c3c60781792"). InnerVolumeSpecName "kube-api-access-d99xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:53:29 crc kubenswrapper[4758]: I1203 18:53:29.951969 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fab9f64-5948-43bd-8af4-4c3c60781792" (UID: "2fab9f64-5948-43bd-8af4-4c3c60781792"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.007023 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d99xp\" (UniqueName: \"kubernetes.io/projected/2fab9f64-5948-43bd-8af4-4c3c60781792-kube-api-access-d99xp\") on node \"crc\" DevicePath \"\"" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.007063 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fab9f64-5948-43bd-8af4-4c3c60781792-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.114414 4758 generic.go:334] "Generic (PLEG): container finished" podID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerID="461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe" exitCode=0 Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.114468 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerDied","Data":"461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe"} Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.114502 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ts8fz" event={"ID":"2fab9f64-5948-43bd-8af4-4c3c60781792","Type":"ContainerDied","Data":"1ad418e10699261d124c28995bafd807e6cd3be269aa68673e5a7778295cb311"} Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.114526 4758 scope.go:117] "RemoveContainer" containerID="461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.114769 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ts8fz" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.137528 4758 scope.go:117] "RemoveContainer" containerID="129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.161708 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ts8fz"] Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.172401 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ts8fz"] Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.188747 4758 scope.go:117] "RemoveContainer" containerID="93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.230649 4758 scope.go:117] "RemoveContainer" containerID="461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe" Dec 03 18:53:30 crc kubenswrapper[4758]: E1203 18:53:30.231088 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe\": container with ID starting with 461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe not found: ID does not exist" containerID="461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.231120 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe"} err="failed to get container status \"461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe\": rpc error: code = NotFound desc = could not find container \"461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe\": container with ID starting with 461540b90bed2369fc28054995a2bd1a835a82f2996e70b50a952d6ccf63b7fe not found: ID does not exist" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.231142 4758 scope.go:117] "RemoveContainer" containerID="129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060" Dec 03 18:53:30 crc kubenswrapper[4758]: E1203 18:53:30.231453 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060\": container with ID starting with 129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060 not found: ID does not exist" containerID="129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.231512 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060"} err="failed to get container status \"129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060\": rpc error: code = NotFound desc = could not find container \"129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060\": container with ID starting with 129abfdb80874b8ad6e8aeac39ad3e7d283be8df954f3568feec66385df8e060 not found: ID does not exist" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.231542 4758 scope.go:117] "RemoveContainer" containerID="93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d" Dec 03 18:53:30 crc kubenswrapper[4758]: E1203 18:53:30.231993 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d\": container with ID starting with 93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d not found: ID does not exist" containerID="93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.232023 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d"} err="failed to get container status \"93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d\": rpc error: code = NotFound desc = could not find container \"93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d\": container with ID starting with 93fcb520ae19c37e037081f89ffe2b3b98bb728ce37412cdaf7771b254f6f32d not found: ID does not exist" Dec 03 18:53:30 crc kubenswrapper[4758]: I1203 18:53:30.282280 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x44mg" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="registry-server" probeResult="failure" output=< Dec 03 18:53:30 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 18:53:30 crc kubenswrapper[4758]: > Dec 03 18:53:31 crc kubenswrapper[4758]: I1203 18:53:31.126464 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" path="/var/lib/kubelet/pods/2fab9f64-5948-43bd-8af4-4c3c60781792/volumes" Dec 03 18:53:39 crc kubenswrapper[4758]: I1203 18:53:39.305842 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:39 crc kubenswrapper[4758]: I1203 18:53:39.353562 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:39 crc kubenswrapper[4758]: I1203 18:53:39.549585 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x44mg"] Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.224986 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x44mg" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="registry-server" containerID="cri-o://350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5" gracePeriod=2 Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.808402 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.991307 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66xjt\" (UniqueName: \"kubernetes.io/projected/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-kube-api-access-66xjt\") pod \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.991508 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-catalog-content\") pod \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.991579 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-utilities\") pod \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\" (UID: \"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972\") " Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.992420 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-utilities" (OuterVolumeSpecName: "utilities") pod "58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" (UID: "58c93d36-9a52-4a2e-96b4-d3bfd2a8d972"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:53:41 crc kubenswrapper[4758]: I1203 18:53:41.998323 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-kube-api-access-66xjt" (OuterVolumeSpecName: "kube-api-access-66xjt") pod "58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" (UID: "58c93d36-9a52-4a2e-96b4-d3bfd2a8d972"). InnerVolumeSpecName "kube-api-access-66xjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.095796 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66xjt\" (UniqueName: \"kubernetes.io/projected/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-kube-api-access-66xjt\") on node \"crc\" DevicePath \"\"" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.096185 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.111197 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" (UID: "58c93d36-9a52-4a2e-96b4-d3bfd2a8d972"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.198869 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.237722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerDied","Data":"350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5"} Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.237798 4758 scope.go:117] "RemoveContainer" containerID="350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.237831 4758 generic.go:334] "Generic (PLEG): container finished" podID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerID="350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5" exitCode=0 Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.237735 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x44mg" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.237882 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x44mg" event={"ID":"58c93d36-9a52-4a2e-96b4-d3bfd2a8d972","Type":"ContainerDied","Data":"2447923717cbcbd40abaf6e1083399eb79c3a549c84979bb04bd01104d556e3a"} Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.262132 4758 scope.go:117] "RemoveContainer" containerID="cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.276183 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x44mg"] Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.286331 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x44mg"] Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.291920 4758 scope.go:117] "RemoveContainer" containerID="c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.347008 4758 scope.go:117] "RemoveContainer" containerID="350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5" Dec 03 18:53:42 crc kubenswrapper[4758]: E1203 18:53:42.350263 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5\": container with ID starting with 350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5 not found: ID does not exist" containerID="350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.350313 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5"} err="failed to get container status \"350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5\": rpc error: code = NotFound desc = could not find container \"350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5\": container with ID starting with 350798709a764b170e6b0150235d806bed0caccc3504dafa891b43a7c99eb0f5 not found: ID does not exist" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.350344 4758 scope.go:117] "RemoveContainer" containerID="cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca" Dec 03 18:53:42 crc kubenswrapper[4758]: E1203 18:53:42.350777 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca\": container with ID starting with cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca not found: ID does not exist" containerID="cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.350832 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca"} err="failed to get container status \"cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca\": rpc error: code = NotFound desc = could not find container \"cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca\": container with ID starting with cc039e73b784caa115aa240f434ba47119e7e95d97453bc7bedc0d32d7a721ca not found: ID does not exist" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.350860 4758 scope.go:117] "RemoveContainer" containerID="c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83" Dec 03 18:53:42 crc kubenswrapper[4758]: E1203 18:53:42.351165 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83\": container with ID starting with c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83 not found: ID does not exist" containerID="c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83" Dec 03 18:53:42 crc kubenswrapper[4758]: I1203 18:53:42.351193 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83"} err="failed to get container status \"c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83\": rpc error: code = NotFound desc = could not find container \"c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83\": container with ID starting with c9a6d79c06b4b7050a36626a1bfcb0ea156a7b872500273f2e3a0ed34d5ede83 not found: ID does not exist" Dec 03 18:53:43 crc kubenswrapper[4758]: I1203 18:53:43.134991 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" path="/var/lib/kubelet/pods/58c93d36-9a52-4a2e-96b4-d3bfd2a8d972/volumes" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.188020 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjjc7"] Dec 03 18:54:00 crc kubenswrapper[4758]: E1203 18:54:00.189309 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="extract-utilities" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189324 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="extract-utilities" Dec 03 18:54:00 crc kubenswrapper[4758]: E1203 18:54:00.189354 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="extract-content" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189361 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="extract-content" Dec 03 18:54:00 crc kubenswrapper[4758]: E1203 18:54:00.189370 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="registry-server" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189376 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="registry-server" Dec 03 18:54:00 crc kubenswrapper[4758]: E1203 18:54:00.189394 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="registry-server" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189399 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="registry-server" Dec 03 18:54:00 crc kubenswrapper[4758]: E1203 18:54:00.189417 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="extract-utilities" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189423 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="extract-utilities" Dec 03 18:54:00 crc kubenswrapper[4758]: E1203 18:54:00.189434 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="extract-content" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189440 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="extract-content" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189630 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c93d36-9a52-4a2e-96b4-d3bfd2a8d972" containerName="registry-server" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.189644 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fab9f64-5948-43bd-8af4-4c3c60781792" containerName="registry-server" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.191187 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.210452 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjjc7"] Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.216887 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-catalog-content\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.217052 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7k8n\" (UniqueName: \"kubernetes.io/projected/a4b2b390-74fd-4211-98f8-cb6e49ecd156-kube-api-access-s7k8n\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.217155 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-utilities\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.319500 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-catalog-content\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.319559 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7k8n\" (UniqueName: \"kubernetes.io/projected/a4b2b390-74fd-4211-98f8-cb6e49ecd156-kube-api-access-s7k8n\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.319653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-utilities\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.320043 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-catalog-content\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.320071 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-utilities\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.339118 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7k8n\" (UniqueName: \"kubernetes.io/projected/a4b2b390-74fd-4211-98f8-cb6e49ecd156-kube-api-access-s7k8n\") pod \"certified-operators-cjjc7\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:00 crc kubenswrapper[4758]: I1203 18:54:00.533600 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:01 crc kubenswrapper[4758]: I1203 18:54:01.092128 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjjc7"] Dec 03 18:54:01 crc kubenswrapper[4758]: I1203 18:54:01.458710 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerID="91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a" exitCode=0 Dec 03 18:54:01 crc kubenswrapper[4758]: I1203 18:54:01.458808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerDied","Data":"91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a"} Dec 03 18:54:01 crc kubenswrapper[4758]: I1203 18:54:01.460879 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerStarted","Data":"8ca365cda4ed394bbdbbe932000f077ea849c04316ea1147ab82941c4174fa99"} Dec 03 18:54:02 crc kubenswrapper[4758]: I1203 18:54:02.481009 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerStarted","Data":"fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081"} Dec 03 18:54:03 crc kubenswrapper[4758]: I1203 18:54:03.507410 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerID="fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081" exitCode=0 Dec 03 18:54:03 crc kubenswrapper[4758]: I1203 18:54:03.507505 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerDied","Data":"fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081"} Dec 03 18:54:04 crc kubenswrapper[4758]: I1203 18:54:04.521436 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerStarted","Data":"67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36"} Dec 03 18:54:04 crc kubenswrapper[4758]: I1203 18:54:04.549460 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjjc7" podStartSLOduration=2.072955709 podStartE2EDuration="4.549435478s" podCreationTimestamp="2025-12-03 18:54:00 +0000 UTC" firstStartedPulling="2025-12-03 18:54:01.460476921 +0000 UTC m=+7096.661853792" lastFinishedPulling="2025-12-03 18:54:03.93695669 +0000 UTC m=+7099.138333561" observedRunningTime="2025-12-03 18:54:04.541609286 +0000 UTC m=+7099.742986157" watchObservedRunningTime="2025-12-03 18:54:04.549435478 +0000 UTC m=+7099.750812339" Dec 03 18:54:10 crc kubenswrapper[4758]: I1203 18:54:10.534572 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:10 crc kubenswrapper[4758]: I1203 18:54:10.535059 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:10 crc kubenswrapper[4758]: I1203 18:54:10.600922 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:10 crc kubenswrapper[4758]: I1203 18:54:10.671744 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:10 crc kubenswrapper[4758]: I1203 18:54:10.840852 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjjc7"] Dec 03 18:54:12 crc kubenswrapper[4758]: I1203 18:54:12.608347 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjjc7" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="registry-server" containerID="cri-o://67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36" gracePeriod=2 Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.166940 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.290355 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7k8n\" (UniqueName: \"kubernetes.io/projected/a4b2b390-74fd-4211-98f8-cb6e49ecd156-kube-api-access-s7k8n\") pod \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.290498 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-catalog-content\") pod \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.290568 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-utilities\") pod \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\" (UID: \"a4b2b390-74fd-4211-98f8-cb6e49ecd156\") " Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.291604 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-utilities" (OuterVolumeSpecName: "utilities") pod "a4b2b390-74fd-4211-98f8-cb6e49ecd156" (UID: "a4b2b390-74fd-4211-98f8-cb6e49ecd156"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.300198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4b2b390-74fd-4211-98f8-cb6e49ecd156-kube-api-access-s7k8n" (OuterVolumeSpecName: "kube-api-access-s7k8n") pod "a4b2b390-74fd-4211-98f8-cb6e49ecd156" (UID: "a4b2b390-74fd-4211-98f8-cb6e49ecd156"). InnerVolumeSpecName "kube-api-access-s7k8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.342446 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4b2b390-74fd-4211-98f8-cb6e49ecd156" (UID: "a4b2b390-74fd-4211-98f8-cb6e49ecd156"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.394127 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.394180 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b2b390-74fd-4211-98f8-cb6e49ecd156-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.394199 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7k8n\" (UniqueName: \"kubernetes.io/projected/a4b2b390-74fd-4211-98f8-cb6e49ecd156-kube-api-access-s7k8n\") on node \"crc\" DevicePath \"\"" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.624999 4758 generic.go:334] "Generic (PLEG): container finished" podID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerID="67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36" exitCode=0 Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.625068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerDied","Data":"67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36"} Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.625118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjjc7" event={"ID":"a4b2b390-74fd-4211-98f8-cb6e49ecd156","Type":"ContainerDied","Data":"8ca365cda4ed394bbdbbe932000f077ea849c04316ea1147ab82941c4174fa99"} Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.625121 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjjc7" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.625150 4758 scope.go:117] "RemoveContainer" containerID="67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.660186 4758 scope.go:117] "RemoveContainer" containerID="fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.681566 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjjc7"] Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.691530 4758 scope.go:117] "RemoveContainer" containerID="91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.695068 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjjc7"] Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.758114 4758 scope.go:117] "RemoveContainer" containerID="67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36" Dec 03 18:54:13 crc kubenswrapper[4758]: E1203 18:54:13.758977 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36\": container with ID starting with 67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36 not found: ID does not exist" containerID="67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.759015 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36"} err="failed to get container status \"67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36\": rpc error: code = NotFound desc = could not find container \"67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36\": container with ID starting with 67b73ec950daa33ea788ff1f554c38cc15088b5ced4068c711774eb3ca6e0d36 not found: ID does not exist" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.759042 4758 scope.go:117] "RemoveContainer" containerID="fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081" Dec 03 18:54:13 crc kubenswrapper[4758]: E1203 18:54:13.759938 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081\": container with ID starting with fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081 not found: ID does not exist" containerID="fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.759963 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081"} err="failed to get container status \"fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081\": rpc error: code = NotFound desc = could not find container \"fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081\": container with ID starting with fe3c0a9d636405e64af8967263c8c79ef88f32c3e173ba646ad828a791786081 not found: ID does not exist" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.759980 4758 scope.go:117] "RemoveContainer" containerID="91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a" Dec 03 18:54:13 crc kubenswrapper[4758]: E1203 18:54:13.760436 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a\": container with ID starting with 91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a not found: ID does not exist" containerID="91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a" Dec 03 18:54:13 crc kubenswrapper[4758]: I1203 18:54:13.760501 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a"} err="failed to get container status \"91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a\": rpc error: code = NotFound desc = could not find container \"91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a\": container with ID starting with 91955084a18b21bfe592995d5e37334a0c7b2f89e0a19c7b8553a68d8f4da27a not found: ID does not exist" Dec 03 18:54:15 crc kubenswrapper[4758]: I1203 18:54:15.132056 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" path="/var/lib/kubelet/pods/a4b2b390-74fd-4211-98f8-cb6e49ecd156/volumes" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.120262 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sjxtf"] Dec 03 18:54:20 crc kubenswrapper[4758]: E1203 18:54:20.121100 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="extract-content" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.121115 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="extract-content" Dec 03 18:54:20 crc kubenswrapper[4758]: E1203 18:54:20.121155 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="extract-utilities" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.121161 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="extract-utilities" Dec 03 18:54:20 crc kubenswrapper[4758]: E1203 18:54:20.121177 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="registry-server" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.121184 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="registry-server" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.121382 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4b2b390-74fd-4211-98f8-cb6e49ecd156" containerName="registry-server" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.125877 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.144646 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjxtf"] Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.146330 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-catalog-content\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.146634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-utilities\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.146976 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r7h9\" (UniqueName: \"kubernetes.io/projected/da10cc52-ba5e-4b36-b7b3-dfa95709d279-kube-api-access-5r7h9\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.249210 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-catalog-content\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.249504 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-utilities\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.249734 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r7h9\" (UniqueName: \"kubernetes.io/projected/da10cc52-ba5e-4b36-b7b3-dfa95709d279-kube-api-access-5r7h9\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.249803 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-catalog-content\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.249959 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-utilities\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.271426 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r7h9\" (UniqueName: \"kubernetes.io/projected/da10cc52-ba5e-4b36-b7b3-dfa95709d279-kube-api-access-5r7h9\") pod \"redhat-marketplace-sjxtf\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.461846 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:20 crc kubenswrapper[4758]: I1203 18:54:20.998079 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjxtf"] Dec 03 18:54:21 crc kubenswrapper[4758]: I1203 18:54:21.699134 4758 generic.go:334] "Generic (PLEG): container finished" podID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerID="ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff" exitCode=0 Dec 03 18:54:21 crc kubenswrapper[4758]: I1203 18:54:21.699183 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjxtf" event={"ID":"da10cc52-ba5e-4b36-b7b3-dfa95709d279","Type":"ContainerDied","Data":"ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff"} Dec 03 18:54:21 crc kubenswrapper[4758]: I1203 18:54:21.699665 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjxtf" event={"ID":"da10cc52-ba5e-4b36-b7b3-dfa95709d279","Type":"ContainerStarted","Data":"866b85e11bed9bf7d5b21855d2348e79e0386a97820a30e4b04312e2c40e1327"} Dec 03 18:54:23 crc kubenswrapper[4758]: I1203 18:54:23.727284 4758 generic.go:334] "Generic (PLEG): container finished" podID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerID="17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754" exitCode=0 Dec 03 18:54:23 crc kubenswrapper[4758]: I1203 18:54:23.727373 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjxtf" event={"ID":"da10cc52-ba5e-4b36-b7b3-dfa95709d279","Type":"ContainerDied","Data":"17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754"} Dec 03 18:54:25 crc kubenswrapper[4758]: I1203 18:54:25.768266 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjxtf" event={"ID":"da10cc52-ba5e-4b36-b7b3-dfa95709d279","Type":"ContainerStarted","Data":"4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758"} Dec 03 18:54:25 crc kubenswrapper[4758]: I1203 18:54:25.794470 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sjxtf" podStartSLOduration=2.858066186 podStartE2EDuration="5.794451548s" podCreationTimestamp="2025-12-03 18:54:20 +0000 UTC" firstStartedPulling="2025-12-03 18:54:21.701805197 +0000 UTC m=+7116.903182058" lastFinishedPulling="2025-12-03 18:54:24.638190569 +0000 UTC m=+7119.839567420" observedRunningTime="2025-12-03 18:54:25.788537338 +0000 UTC m=+7120.989914199" watchObservedRunningTime="2025-12-03 18:54:25.794451548 +0000 UTC m=+7120.995828409" Dec 03 18:54:30 crc kubenswrapper[4758]: I1203 18:54:30.463059 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:30 crc kubenswrapper[4758]: I1203 18:54:30.465244 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:30 crc kubenswrapper[4758]: I1203 18:54:30.516734 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:30 crc kubenswrapper[4758]: I1203 18:54:30.883005 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:31 crc kubenswrapper[4758]: I1203 18:54:31.387905 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjxtf"] Dec 03 18:54:32 crc kubenswrapper[4758]: I1203 18:54:32.842063 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sjxtf" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="registry-server" containerID="cri-o://4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758" gracePeriod=2 Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.349985 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.474820 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-catalog-content\") pod \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.475016 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r7h9\" (UniqueName: \"kubernetes.io/projected/da10cc52-ba5e-4b36-b7b3-dfa95709d279-kube-api-access-5r7h9\") pod \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.475308 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-utilities\") pod \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\" (UID: \"da10cc52-ba5e-4b36-b7b3-dfa95709d279\") " Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.476510 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-utilities" (OuterVolumeSpecName: "utilities") pod "da10cc52-ba5e-4b36-b7b3-dfa95709d279" (UID: "da10cc52-ba5e-4b36-b7b3-dfa95709d279"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.485446 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da10cc52-ba5e-4b36-b7b3-dfa95709d279-kube-api-access-5r7h9" (OuterVolumeSpecName: "kube-api-access-5r7h9") pod "da10cc52-ba5e-4b36-b7b3-dfa95709d279" (UID: "da10cc52-ba5e-4b36-b7b3-dfa95709d279"). InnerVolumeSpecName "kube-api-access-5r7h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.499963 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da10cc52-ba5e-4b36-b7b3-dfa95709d279" (UID: "da10cc52-ba5e-4b36-b7b3-dfa95709d279"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.618166 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r7h9\" (UniqueName: \"kubernetes.io/projected/da10cc52-ba5e-4b36-b7b3-dfa95709d279-kube-api-access-5r7h9\") on node \"crc\" DevicePath \"\"" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.618212 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.618224 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da10cc52-ba5e-4b36-b7b3-dfa95709d279-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.855324 4758 generic.go:334] "Generic (PLEG): container finished" podID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerID="4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758" exitCode=0 Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.855384 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjxtf" event={"ID":"da10cc52-ba5e-4b36-b7b3-dfa95709d279","Type":"ContainerDied","Data":"4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758"} Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.855462 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjxtf" event={"ID":"da10cc52-ba5e-4b36-b7b3-dfa95709d279","Type":"ContainerDied","Data":"866b85e11bed9bf7d5b21855d2348e79e0386a97820a30e4b04312e2c40e1327"} Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.855483 4758 scope.go:117] "RemoveContainer" containerID="4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.855398 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sjxtf" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.876833 4758 scope.go:117] "RemoveContainer" containerID="17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.897425 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjxtf"] Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.906393 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjxtf"] Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.912359 4758 scope.go:117] "RemoveContainer" containerID="ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.946288 4758 scope.go:117] "RemoveContainer" containerID="4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758" Dec 03 18:54:33 crc kubenswrapper[4758]: E1203 18:54:33.946745 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758\": container with ID starting with 4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758 not found: ID does not exist" containerID="4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.946814 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758"} err="failed to get container status \"4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758\": rpc error: code = NotFound desc = could not find container \"4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758\": container with ID starting with 4694c441a13d37354e5b008d8a0a4042d0b2655cb79c3b2a598753fe6cc43758 not found: ID does not exist" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.946869 4758 scope.go:117] "RemoveContainer" containerID="17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754" Dec 03 18:54:33 crc kubenswrapper[4758]: E1203 18:54:33.947227 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754\": container with ID starting with 17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754 not found: ID does not exist" containerID="17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.947256 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754"} err="failed to get container status \"17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754\": rpc error: code = NotFound desc = could not find container \"17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754\": container with ID starting with 17f1c1a1c50db27de11d75ccfde6dde6121c04b97cf6e7719001b70c753c8754 not found: ID does not exist" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.947277 4758 scope.go:117] "RemoveContainer" containerID="ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff" Dec 03 18:54:33 crc kubenswrapper[4758]: E1203 18:54:33.947551 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff\": container with ID starting with ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff not found: ID does not exist" containerID="ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff" Dec 03 18:54:33 crc kubenswrapper[4758]: I1203 18:54:33.947574 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff"} err="failed to get container status \"ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff\": rpc error: code = NotFound desc = could not find container \"ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff\": container with ID starting with ee41817abe2796f0236cd0898bd309a69c883b7e7a5abf3da017039366990dff not found: ID does not exist" Dec 03 18:54:35 crc kubenswrapper[4758]: I1203 18:54:35.126737 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" path="/var/lib/kubelet/pods/da10cc52-ba5e-4b36-b7b3-dfa95709d279/volumes" Dec 03 18:54:41 crc kubenswrapper[4758]: I1203 18:54:41.394190 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:54:41 crc kubenswrapper[4758]: I1203 18:54:41.394915 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:55:07 crc kubenswrapper[4758]: I1203 18:55:07.053424 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-d98a-account-create-update-r98bs"] Dec 03 18:55:07 crc kubenswrapper[4758]: I1203 18:55:07.067987 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-m5jpt"] Dec 03 18:55:07 crc kubenswrapper[4758]: I1203 18:55:07.078932 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-d98a-account-create-update-r98bs"] Dec 03 18:55:07 crc kubenswrapper[4758]: I1203 18:55:07.089564 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-m5jpt"] Dec 03 18:55:07 crc kubenswrapper[4758]: I1203 18:55:07.126765 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae" path="/var/lib/kubelet/pods/9f7e58a7-b051-4bd5-a0b5-0cd2fe38ccae/volumes" Dec 03 18:55:07 crc kubenswrapper[4758]: I1203 18:55:07.128912 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f803c48e-75e8-45a6-9d60-28961db451ae" path="/var/lib/kubelet/pods/f803c48e-75e8-45a6-9d60-28961db451ae/volumes" Dec 03 18:55:11 crc kubenswrapper[4758]: I1203 18:55:11.397306 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:55:11 crc kubenswrapper[4758]: I1203 18:55:11.397779 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:55:16 crc kubenswrapper[4758]: I1203 18:55:16.663548 4758 scope.go:117] "RemoveContainer" containerID="33dd7c7a17f3cc077f4270dce9b552b5ade7f75811bf8f0260cac244d39aa7be" Dec 03 18:55:16 crc kubenswrapper[4758]: I1203 18:55:16.696322 4758 scope.go:117] "RemoveContainer" containerID="a7fe3871ab507ceceb3f54e5764fd99385d76ba7abdb49c745d88c109c88fb58" Dec 03 18:55:19 crc kubenswrapper[4758]: I1203 18:55:19.060973 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-nh66c"] Dec 03 18:55:19 crc kubenswrapper[4758]: I1203 18:55:19.082711 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-nh66c"] Dec 03 18:55:19 crc kubenswrapper[4758]: I1203 18:55:19.127169 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2be426c3-873e-4472-8efc-abd754d826aa" path="/var/lib/kubelet/pods/2be426c3-873e-4472-8efc-abd754d826aa/volumes" Dec 03 18:55:40 crc kubenswrapper[4758]: I1203 18:55:40.033952 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-5fqlp"] Dec 03 18:55:40 crc kubenswrapper[4758]: I1203 18:55:40.044355 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-bc50-account-create-update-gmnkc"] Dec 03 18:55:40 crc kubenswrapper[4758]: I1203 18:55:40.054375 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-5fqlp"] Dec 03 18:55:40 crc kubenswrapper[4758]: I1203 18:55:40.062388 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-bc50-account-create-update-gmnkc"] Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.131382 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b94047-0675-47ef-bd70-bffe0d411da6" path="/var/lib/kubelet/pods/62b94047-0675-47ef-bd70-bffe0d411da6/volumes" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.132135 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99f6f035-7ce0-458e-ac7d-10767462400d" path="/var/lib/kubelet/pods/99f6f035-7ce0-458e-ac7d-10767462400d/volumes" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.394834 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.395146 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.395182 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.395980 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.396030 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" gracePeriod=600 Dec 03 18:55:41 crc kubenswrapper[4758]: E1203 18:55:41.529079 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.588371 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" exitCode=0 Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.588417 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74"} Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.588454 4758 scope.go:117] "RemoveContainer" containerID="64411234ffdaf6687b4f6e58c921ca9496d13f2f2e70093e2409ece7a93f666b" Dec 03 18:55:41 crc kubenswrapper[4758]: I1203 18:55:41.589175 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:55:41 crc kubenswrapper[4758]: E1203 18:55:41.589487 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:55:53 crc kubenswrapper[4758]: I1203 18:55:53.080394 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-2bngz"] Dec 03 18:55:53 crc kubenswrapper[4758]: I1203 18:55:53.088974 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-2bngz"] Dec 03 18:55:53 crc kubenswrapper[4758]: I1203 18:55:53.115073 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:55:53 crc kubenswrapper[4758]: E1203 18:55:53.115610 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:55:53 crc kubenswrapper[4758]: I1203 18:55:53.132140 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0" path="/var/lib/kubelet/pods/3a2b2cf3-4c51-40d3-a274-01f3cb9a73b0/volumes" Dec 03 18:56:05 crc kubenswrapper[4758]: I1203 18:56:05.121065 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:56:05 crc kubenswrapper[4758]: E1203 18:56:05.121894 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:56:16 crc kubenswrapper[4758]: I1203 18:56:16.886353 4758 scope.go:117] "RemoveContainer" containerID="7580f0ffaeb2004e4cf8e0068e133822d1042b9b129f2a507420abb145b49524" Dec 03 18:56:16 crc kubenswrapper[4758]: I1203 18:56:16.919270 4758 scope.go:117] "RemoveContainer" containerID="6b459a65bff49fbcef0d38c12fe44a34c36773cafd8dff8903a96ed535ecc7e6" Dec 03 18:56:16 crc kubenswrapper[4758]: I1203 18:56:16.989840 4758 scope.go:117] "RemoveContainer" containerID="0cfbc7b94a728c199a84cff2f2265f566e20f88313dc4dd78cd6404cee1d495f" Dec 03 18:56:17 crc kubenswrapper[4758]: I1203 18:56:17.047974 4758 scope.go:117] "RemoveContainer" containerID="23a1c883b688c374f4c94115237ebf5d253db8254f8c276989a615a0d01496d7" Dec 03 18:56:20 crc kubenswrapper[4758]: I1203 18:56:20.114774 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:56:20 crc kubenswrapper[4758]: E1203 18:56:20.115599 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:56:33 crc kubenswrapper[4758]: I1203 18:56:33.115405 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:56:33 crc kubenswrapper[4758]: E1203 18:56:33.116585 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:56:44 crc kubenswrapper[4758]: I1203 18:56:44.115919 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:56:44 crc kubenswrapper[4758]: E1203 18:56:44.116824 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:56:55 crc kubenswrapper[4758]: I1203 18:56:55.121026 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:56:55 crc kubenswrapper[4758]: E1203 18:56:55.122069 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:57:09 crc kubenswrapper[4758]: I1203 18:57:09.115877 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:57:09 crc kubenswrapper[4758]: E1203 18:57:09.116869 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:57:24 crc kubenswrapper[4758]: I1203 18:57:24.116220 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:57:24 crc kubenswrapper[4758]: E1203 18:57:24.117413 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:57:39 crc kubenswrapper[4758]: I1203 18:57:39.114415 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:57:39 crc kubenswrapper[4758]: E1203 18:57:39.115317 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:57:53 crc kubenswrapper[4758]: I1203 18:57:53.114888 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:57:53 crc kubenswrapper[4758]: E1203 18:57:53.115910 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:58:05 crc kubenswrapper[4758]: I1203 18:58:05.123670 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:58:05 crc kubenswrapper[4758]: E1203 18:58:05.124520 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:58:20 crc kubenswrapper[4758]: I1203 18:58:20.114640 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:58:20 crc kubenswrapper[4758]: E1203 18:58:20.117244 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:58:31 crc kubenswrapper[4758]: I1203 18:58:31.114873 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:58:31 crc kubenswrapper[4758]: E1203 18:58:31.115637 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:58:45 crc kubenswrapper[4758]: I1203 18:58:45.136113 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:58:45 crc kubenswrapper[4758]: E1203 18:58:45.137359 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:58:57 crc kubenswrapper[4758]: I1203 18:58:57.115138 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:58:57 crc kubenswrapper[4758]: E1203 18:58:57.116283 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:59:12 crc kubenswrapper[4758]: I1203 18:59:12.114910 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:59:12 crc kubenswrapper[4758]: E1203 18:59:12.115770 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:59:17 crc kubenswrapper[4758]: I1203 18:59:17.116198 4758 generic.go:334] "Generic (PLEG): container finished" podID="650575e1-2ffe-41e4-887a-ae4e451bc53a" containerID="c1abc6ce129c0e89d9c1a1c461158862c0cc032849fb1867590ee158d2adf075" exitCode=0 Dec 03 18:59:17 crc kubenswrapper[4758]: I1203 18:59:17.134169 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" event={"ID":"650575e1-2ffe-41e4-887a-ae4e451bc53a","Type":"ContainerDied","Data":"c1abc6ce129c0e89d9c1a1c461158862c0cc032849fb1867590ee158d2adf075"} Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.696956 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.852900 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ceph\") pod \"650575e1-2ffe-41e4-887a-ae4e451bc53a\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.852959 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-inventory\") pod \"650575e1-2ffe-41e4-887a-ae4e451bc53a\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.853036 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-tripleo-cleanup-combined-ca-bundle\") pod \"650575e1-2ffe-41e4-887a-ae4e451bc53a\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.853056 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxxft\" (UniqueName: \"kubernetes.io/projected/650575e1-2ffe-41e4-887a-ae4e451bc53a-kube-api-access-qxxft\") pod \"650575e1-2ffe-41e4-887a-ae4e451bc53a\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.853794 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ssh-key\") pod \"650575e1-2ffe-41e4-887a-ae4e451bc53a\" (UID: \"650575e1-2ffe-41e4-887a-ae4e451bc53a\") " Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.858939 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "650575e1-2ffe-41e4-887a-ae4e451bc53a" (UID: "650575e1-2ffe-41e4-887a-ae4e451bc53a"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.859969 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650575e1-2ffe-41e4-887a-ae4e451bc53a-kube-api-access-qxxft" (OuterVolumeSpecName: "kube-api-access-qxxft") pod "650575e1-2ffe-41e4-887a-ae4e451bc53a" (UID: "650575e1-2ffe-41e4-887a-ae4e451bc53a"). InnerVolumeSpecName "kube-api-access-qxxft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.867261 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ceph" (OuterVolumeSpecName: "ceph") pod "650575e1-2ffe-41e4-887a-ae4e451bc53a" (UID: "650575e1-2ffe-41e4-887a-ae4e451bc53a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.884393 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-inventory" (OuterVolumeSpecName: "inventory") pod "650575e1-2ffe-41e4-887a-ae4e451bc53a" (UID: "650575e1-2ffe-41e4-887a-ae4e451bc53a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.897542 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "650575e1-2ffe-41e4-887a-ae4e451bc53a" (UID: "650575e1-2ffe-41e4-887a-ae4e451bc53a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.956660 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.956698 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.956710 4758 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.956723 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxxft\" (UniqueName: \"kubernetes.io/projected/650575e1-2ffe-41e4-887a-ae4e451bc53a-kube-api-access-qxxft\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:18 crc kubenswrapper[4758]: I1203 18:59:18.956732 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/650575e1-2ffe-41e4-887a-ae4e451bc53a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:19 crc kubenswrapper[4758]: I1203 18:59:19.144315 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" event={"ID":"650575e1-2ffe-41e4-887a-ae4e451bc53a","Type":"ContainerDied","Data":"b7fcf5e6edc1dcef8477d35ff29bc61d310c13a0674fd405f0718b02346b54df"} Dec 03 18:59:19 crc kubenswrapper[4758]: I1203 18:59:19.144384 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7fcf5e6edc1dcef8477d35ff29bc61d310c13a0674fd405f0718b02346b54df" Dec 03 18:59:19 crc kubenswrapper[4758]: I1203 18:59:19.144400 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.115352 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:59:23 crc kubenswrapper[4758]: E1203 18:59:23.116713 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.345828 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-jwtcm"] Dec 03 18:59:23 crc kubenswrapper[4758]: E1203 18:59:23.346762 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="extract-utilities" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.346910 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="extract-utilities" Dec 03 18:59:23 crc kubenswrapper[4758]: E1203 18:59:23.347056 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="extract-content" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.347172 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="extract-content" Dec 03 18:59:23 crc kubenswrapper[4758]: E1203 18:59:23.347542 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650575e1-2ffe-41e4-887a-ae4e451bc53a" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.347713 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="650575e1-2ffe-41e4-887a-ae4e451bc53a" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 03 18:59:23 crc kubenswrapper[4758]: E1203 18:59:23.347845 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="registry-server" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.347959 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="registry-server" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.348447 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="da10cc52-ba5e-4b36-b7b3-dfa95709d279" containerName="registry-server" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.348631 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="650575e1-2ffe-41e4-887a-ae4e451bc53a" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.350102 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.355281 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.355599 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.355945 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.356283 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.361163 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-jwtcm"] Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.507003 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.507099 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.508457 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgsgs\" (UniqueName: \"kubernetes.io/projected/cbe9768c-6273-4788-8aa7-8ed927817cc5-kube-api-access-sgsgs\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.508567 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ceph\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.508676 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-inventory\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.611458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgsgs\" (UniqueName: \"kubernetes.io/projected/cbe9768c-6273-4788-8aa7-8ed927817cc5-kube-api-access-sgsgs\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.611623 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ceph\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.611787 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-inventory\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.611922 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.612009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.617504 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.618607 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.623156 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ceph\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.625888 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-inventory\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.628983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgsgs\" (UniqueName: \"kubernetes.io/projected/cbe9768c-6273-4788-8aa7-8ed927817cc5-kube-api-access-sgsgs\") pod \"bootstrap-openstack-openstack-cell1-jwtcm\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:23 crc kubenswrapper[4758]: I1203 18:59:23.727468 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 18:59:24 crc kubenswrapper[4758]: I1203 18:59:24.345496 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:59:24 crc kubenswrapper[4758]: I1203 18:59:24.349388 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-jwtcm"] Dec 03 18:59:25 crc kubenswrapper[4758]: I1203 18:59:25.216373 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" event={"ID":"cbe9768c-6273-4788-8aa7-8ed927817cc5","Type":"ContainerStarted","Data":"c5a4b2d170c69b1fc701bc2d35bb90793160802c341aefd7a61cd34506628a14"} Dec 03 18:59:25 crc kubenswrapper[4758]: I1203 18:59:25.217000 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" event={"ID":"cbe9768c-6273-4788-8aa7-8ed927817cc5","Type":"ContainerStarted","Data":"3c70a7a19d611b2112a1b7217375210f4c2b99621021e615c9aaa2fb739e7343"} Dec 03 18:59:25 crc kubenswrapper[4758]: I1203 18:59:25.261915 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" podStartSLOduration=2.04454483 podStartE2EDuration="2.261891995s" podCreationTimestamp="2025-12-03 18:59:23 +0000 UTC" firstStartedPulling="2025-12-03 18:59:24.345314967 +0000 UTC m=+7419.546691828" lastFinishedPulling="2025-12-03 18:59:24.562662132 +0000 UTC m=+7419.764038993" observedRunningTime="2025-12-03 18:59:25.248266257 +0000 UTC m=+7420.449643168" watchObservedRunningTime="2025-12-03 18:59:25.261891995 +0000 UTC m=+7420.463268866" Dec 03 18:59:35 crc kubenswrapper[4758]: I1203 18:59:35.130994 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:59:35 crc kubenswrapper[4758]: E1203 18:59:35.131760 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:59:47 crc kubenswrapper[4758]: I1203 18:59:47.114914 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:59:47 crc kubenswrapper[4758]: E1203 18:59:47.115741 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 18:59:58 crc kubenswrapper[4758]: I1203 18:59:58.115199 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 18:59:58 crc kubenswrapper[4758]: E1203 18:59:58.116164 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.203787 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9"] Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.207369 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.210856 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.213606 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.220988 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9"] Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.312169 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n6gg\" (UniqueName: \"kubernetes.io/projected/03d7da42-ab46-4ef8-aede-79f3489ed18c-kube-api-access-8n6gg\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.312492 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d7da42-ab46-4ef8-aede-79f3489ed18c-secret-volume\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.312613 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d7da42-ab46-4ef8-aede-79f3489ed18c-config-volume\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.414569 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d7da42-ab46-4ef8-aede-79f3489ed18c-config-volume\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.414885 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n6gg\" (UniqueName: \"kubernetes.io/projected/03d7da42-ab46-4ef8-aede-79f3489ed18c-kube-api-access-8n6gg\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.415008 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d7da42-ab46-4ef8-aede-79f3489ed18c-secret-volume\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.416540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d7da42-ab46-4ef8-aede-79f3489ed18c-config-volume\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.428251 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d7da42-ab46-4ef8-aede-79f3489ed18c-secret-volume\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.444934 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n6gg\" (UniqueName: \"kubernetes.io/projected/03d7da42-ab46-4ef8-aede-79f3489ed18c-kube-api-access-8n6gg\") pod \"collect-profiles-29413140-45tx9\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:00 crc kubenswrapper[4758]: I1203 19:00:00.550789 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:01 crc kubenswrapper[4758]: I1203 19:00:01.038044 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9"] Dec 03 19:00:01 crc kubenswrapper[4758]: W1203 19:00:01.049096 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03d7da42_ab46_4ef8_aede_79f3489ed18c.slice/crio-722f11a083f0d1755e40877b77c779a86d1cf632c82a44320518603a7d43689c WatchSource:0}: Error finding container 722f11a083f0d1755e40877b77c779a86d1cf632c82a44320518603a7d43689c: Status 404 returned error can't find the container with id 722f11a083f0d1755e40877b77c779a86d1cf632c82a44320518603a7d43689c Dec 03 19:00:01 crc kubenswrapper[4758]: I1203 19:00:01.699247 4758 generic.go:334] "Generic (PLEG): container finished" podID="03d7da42-ab46-4ef8-aede-79f3489ed18c" containerID="eed8eba36e8b27534e82423aabbe66e5ff3f5db85bc22fca0e643ccce51a59f0" exitCode=0 Dec 03 19:00:01 crc kubenswrapper[4758]: I1203 19:00:01.699354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" event={"ID":"03d7da42-ab46-4ef8-aede-79f3489ed18c","Type":"ContainerDied","Data":"eed8eba36e8b27534e82423aabbe66e5ff3f5db85bc22fca0e643ccce51a59f0"} Dec 03 19:00:01 crc kubenswrapper[4758]: I1203 19:00:01.699745 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" event={"ID":"03d7da42-ab46-4ef8-aede-79f3489ed18c","Type":"ContainerStarted","Data":"722f11a083f0d1755e40877b77c779a86d1cf632c82a44320518603a7d43689c"} Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.093889 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.175781 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d7da42-ab46-4ef8-aede-79f3489ed18c-secret-volume\") pod \"03d7da42-ab46-4ef8-aede-79f3489ed18c\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.176004 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d7da42-ab46-4ef8-aede-79f3489ed18c-config-volume\") pod \"03d7da42-ab46-4ef8-aede-79f3489ed18c\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.176187 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n6gg\" (UniqueName: \"kubernetes.io/projected/03d7da42-ab46-4ef8-aede-79f3489ed18c-kube-api-access-8n6gg\") pod \"03d7da42-ab46-4ef8-aede-79f3489ed18c\" (UID: \"03d7da42-ab46-4ef8-aede-79f3489ed18c\") " Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.177767 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d7da42-ab46-4ef8-aede-79f3489ed18c-config-volume" (OuterVolumeSpecName: "config-volume") pod "03d7da42-ab46-4ef8-aede-79f3489ed18c" (UID: "03d7da42-ab46-4ef8-aede-79f3489ed18c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.186431 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d7da42-ab46-4ef8-aede-79f3489ed18c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "03d7da42-ab46-4ef8-aede-79f3489ed18c" (UID: "03d7da42-ab46-4ef8-aede-79f3489ed18c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.186508 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d7da42-ab46-4ef8-aede-79f3489ed18c-kube-api-access-8n6gg" (OuterVolumeSpecName: "kube-api-access-8n6gg") pod "03d7da42-ab46-4ef8-aede-79f3489ed18c" (UID: "03d7da42-ab46-4ef8-aede-79f3489ed18c"). InnerVolumeSpecName "kube-api-access-8n6gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.278844 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d7da42-ab46-4ef8-aede-79f3489ed18c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.278875 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d7da42-ab46-4ef8-aede-79f3489ed18c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.278884 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n6gg\" (UniqueName: \"kubernetes.io/projected/03d7da42-ab46-4ef8-aede-79f3489ed18c-kube-api-access-8n6gg\") on node \"crc\" DevicePath \"\"" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.719146 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" event={"ID":"03d7da42-ab46-4ef8-aede-79f3489ed18c","Type":"ContainerDied","Data":"722f11a083f0d1755e40877b77c779a86d1cf632c82a44320518603a7d43689c"} Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.719495 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="722f11a083f0d1755e40877b77c779a86d1cf632c82a44320518603a7d43689c" Dec 03 19:00:03 crc kubenswrapper[4758]: I1203 19:00:03.719190 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9" Dec 03 19:00:04 crc kubenswrapper[4758]: I1203 19:00:04.185478 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz"] Dec 03 19:00:04 crc kubenswrapper[4758]: I1203 19:00:04.203126 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-6w4pz"] Dec 03 19:00:05 crc kubenswrapper[4758]: I1203 19:00:05.129276 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de23699d-45ea-4220-8b62-6214fd0f33b5" path="/var/lib/kubelet/pods/de23699d-45ea-4220-8b62-6214fd0f33b5/volumes" Dec 03 19:00:13 crc kubenswrapper[4758]: I1203 19:00:13.114919 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 19:00:13 crc kubenswrapper[4758]: E1203 19:00:13.116032 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:00:17 crc kubenswrapper[4758]: I1203 19:00:17.253597 4758 scope.go:117] "RemoveContainer" containerID="0d07c9f24299f0fe2ec54708ac70d5cd823422490cb1e05af0c98d7e218cd143" Dec 03 19:00:27 crc kubenswrapper[4758]: I1203 19:00:27.115003 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 19:00:27 crc kubenswrapper[4758]: E1203 19:00:27.115885 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:00:39 crc kubenswrapper[4758]: I1203 19:00:39.115114 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 19:00:39 crc kubenswrapper[4758]: E1203 19:00:39.115911 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:00:54 crc kubenswrapper[4758]: I1203 19:00:54.114853 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 19:00:55 crc kubenswrapper[4758]: I1203 19:00:55.250864 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"27aab133a3e8862abe59f22c91cb6fa989efdafc746712ea304043f43c4083a5"} Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.166766 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413141-k86xb"] Dec 03 19:01:00 crc kubenswrapper[4758]: E1203 19:01:00.168062 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d7da42-ab46-4ef8-aede-79f3489ed18c" containerName="collect-profiles" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.168087 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d7da42-ab46-4ef8-aede-79f3489ed18c" containerName="collect-profiles" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.168460 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d7da42-ab46-4ef8-aede-79f3489ed18c" containerName="collect-profiles" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.169769 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.180631 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413141-k86xb"] Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.323015 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-fernet-keys\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.323104 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8glb\" (UniqueName: \"kubernetes.io/projected/1ee10a9b-0fcf-4701-802d-480e6131feed-kube-api-access-v8glb\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.323201 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-combined-ca-bundle\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.323260 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-config-data\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.425328 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-combined-ca-bundle\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.425403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-config-data\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.425515 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-fernet-keys\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.425577 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8glb\" (UniqueName: \"kubernetes.io/projected/1ee10a9b-0fcf-4701-802d-480e6131feed-kube-api-access-v8glb\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.432582 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-fernet-keys\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.438836 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-combined-ca-bundle\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.442246 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-config-data\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.445389 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8glb\" (UniqueName: \"kubernetes.io/projected/1ee10a9b-0fcf-4701-802d-480e6131feed-kube-api-access-v8glb\") pod \"keystone-cron-29413141-k86xb\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:00 crc kubenswrapper[4758]: I1203 19:01:00.515761 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:01 crc kubenswrapper[4758]: I1203 19:01:01.042440 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413141-k86xb"] Dec 03 19:01:01 crc kubenswrapper[4758]: W1203 19:01:01.056151 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ee10a9b_0fcf_4701_802d_480e6131feed.slice/crio-4b1c206d72fb9b3e63363311bbd217ebe8c66f308155e5f0df2cb0c81f21b44e WatchSource:0}: Error finding container 4b1c206d72fb9b3e63363311bbd217ebe8c66f308155e5f0df2cb0c81f21b44e: Status 404 returned error can't find the container with id 4b1c206d72fb9b3e63363311bbd217ebe8c66f308155e5f0df2cb0c81f21b44e Dec 03 19:01:01 crc kubenswrapper[4758]: I1203 19:01:01.329032 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-k86xb" event={"ID":"1ee10a9b-0fcf-4701-802d-480e6131feed","Type":"ContainerStarted","Data":"4b1c206d72fb9b3e63363311bbd217ebe8c66f308155e5f0df2cb0c81f21b44e"} Dec 03 19:01:02 crc kubenswrapper[4758]: I1203 19:01:02.339550 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-k86xb" event={"ID":"1ee10a9b-0fcf-4701-802d-480e6131feed","Type":"ContainerStarted","Data":"ff860593cedc94dbce1a11ccd8e98d8f5e66797450982c8750e593d5e12eef82"} Dec 03 19:01:06 crc kubenswrapper[4758]: I1203 19:01:06.378882 4758 generic.go:334] "Generic (PLEG): container finished" podID="1ee10a9b-0fcf-4701-802d-480e6131feed" containerID="ff860593cedc94dbce1a11ccd8e98d8f5e66797450982c8750e593d5e12eef82" exitCode=0 Dec 03 19:01:06 crc kubenswrapper[4758]: I1203 19:01:06.378971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-k86xb" event={"ID":"1ee10a9b-0fcf-4701-802d-480e6131feed","Type":"ContainerDied","Data":"ff860593cedc94dbce1a11ccd8e98d8f5e66797450982c8750e593d5e12eef82"} Dec 03 19:01:07 crc kubenswrapper[4758]: I1203 19:01:07.887489 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:07 crc kubenswrapper[4758]: I1203 19:01:07.994561 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8glb\" (UniqueName: \"kubernetes.io/projected/1ee10a9b-0fcf-4701-802d-480e6131feed-kube-api-access-v8glb\") pod \"1ee10a9b-0fcf-4701-802d-480e6131feed\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " Dec 03 19:01:07 crc kubenswrapper[4758]: I1203 19:01:07.994617 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-fernet-keys\") pod \"1ee10a9b-0fcf-4701-802d-480e6131feed\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " Dec 03 19:01:07 crc kubenswrapper[4758]: I1203 19:01:07.994765 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-combined-ca-bundle\") pod \"1ee10a9b-0fcf-4701-802d-480e6131feed\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " Dec 03 19:01:07 crc kubenswrapper[4758]: I1203 19:01:07.994805 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-config-data\") pod \"1ee10a9b-0fcf-4701-802d-480e6131feed\" (UID: \"1ee10a9b-0fcf-4701-802d-480e6131feed\") " Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.001132 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ee10a9b-0fcf-4701-802d-480e6131feed-kube-api-access-v8glb" (OuterVolumeSpecName: "kube-api-access-v8glb") pod "1ee10a9b-0fcf-4701-802d-480e6131feed" (UID: "1ee10a9b-0fcf-4701-802d-480e6131feed"). InnerVolumeSpecName "kube-api-access-v8glb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.001500 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1ee10a9b-0fcf-4701-802d-480e6131feed" (UID: "1ee10a9b-0fcf-4701-802d-480e6131feed"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.027238 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ee10a9b-0fcf-4701-802d-480e6131feed" (UID: "1ee10a9b-0fcf-4701-802d-480e6131feed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.067641 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-config-data" (OuterVolumeSpecName: "config-data") pod "1ee10a9b-0fcf-4701-802d-480e6131feed" (UID: "1ee10a9b-0fcf-4701-802d-480e6131feed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.097530 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.097566 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.097575 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8glb\" (UniqueName: \"kubernetes.io/projected/1ee10a9b-0fcf-4701-802d-480e6131feed-kube-api-access-v8glb\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.097586 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1ee10a9b-0fcf-4701-802d-480e6131feed-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.412027 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-k86xb" Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.411886 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-k86xb" event={"ID":"1ee10a9b-0fcf-4701-802d-480e6131feed","Type":"ContainerDied","Data":"4b1c206d72fb9b3e63363311bbd217ebe8c66f308155e5f0df2cb0c81f21b44e"} Dec 03 19:01:08 crc kubenswrapper[4758]: I1203 19:01:08.412863 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b1c206d72fb9b3e63363311bbd217ebe8c66f308155e5f0df2cb0c81f21b44e" Dec 03 19:02:49 crc kubenswrapper[4758]: I1203 19:02:49.557070 4758 generic.go:334] "Generic (PLEG): container finished" podID="cbe9768c-6273-4788-8aa7-8ed927817cc5" containerID="c5a4b2d170c69b1fc701bc2d35bb90793160802c341aefd7a61cd34506628a14" exitCode=0 Dec 03 19:02:49 crc kubenswrapper[4758]: I1203 19:02:49.557174 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" event={"ID":"cbe9768c-6273-4788-8aa7-8ed927817cc5","Type":"ContainerDied","Data":"c5a4b2d170c69b1fc701bc2d35bb90793160802c341aefd7a61cd34506628a14"} Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.083894 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.209864 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ssh-key\") pod \"cbe9768c-6273-4788-8aa7-8ed927817cc5\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.210171 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-bootstrap-combined-ca-bundle\") pod \"cbe9768c-6273-4788-8aa7-8ed927817cc5\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.210276 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-inventory\") pod \"cbe9768c-6273-4788-8aa7-8ed927817cc5\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.210376 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgsgs\" (UniqueName: \"kubernetes.io/projected/cbe9768c-6273-4788-8aa7-8ed927817cc5-kube-api-access-sgsgs\") pod \"cbe9768c-6273-4788-8aa7-8ed927817cc5\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.210453 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ceph\") pod \"cbe9768c-6273-4788-8aa7-8ed927817cc5\" (UID: \"cbe9768c-6273-4788-8aa7-8ed927817cc5\") " Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.215628 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "cbe9768c-6273-4788-8aa7-8ed927817cc5" (UID: "cbe9768c-6273-4788-8aa7-8ed927817cc5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.215939 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe9768c-6273-4788-8aa7-8ed927817cc5-kube-api-access-sgsgs" (OuterVolumeSpecName: "kube-api-access-sgsgs") pod "cbe9768c-6273-4788-8aa7-8ed927817cc5" (UID: "cbe9768c-6273-4788-8aa7-8ed927817cc5"). InnerVolumeSpecName "kube-api-access-sgsgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.218270 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ceph" (OuterVolumeSpecName: "ceph") pod "cbe9768c-6273-4788-8aa7-8ed927817cc5" (UID: "cbe9768c-6273-4788-8aa7-8ed927817cc5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.258726 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbe9768c-6273-4788-8aa7-8ed927817cc5" (UID: "cbe9768c-6273-4788-8aa7-8ed927817cc5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.270402 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-inventory" (OuterVolumeSpecName: "inventory") pod "cbe9768c-6273-4788-8aa7-8ed927817cc5" (UID: "cbe9768c-6273-4788-8aa7-8ed927817cc5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.313380 4758 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.313440 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.313455 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgsgs\" (UniqueName: \"kubernetes.io/projected/cbe9768c-6273-4788-8aa7-8ed927817cc5-kube-api-access-sgsgs\") on node \"crc\" DevicePath \"\"" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.313467 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.313479 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbe9768c-6273-4788-8aa7-8ed927817cc5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.580476 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.580388 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jwtcm" event={"ID":"cbe9768c-6273-4788-8aa7-8ed927817cc5","Type":"ContainerDied","Data":"3c70a7a19d611b2112a1b7217375210f4c2b99621021e615c9aaa2fb739e7343"} Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.582876 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c70a7a19d611b2112a1b7217375210f4c2b99621021e615c9aaa2fb739e7343" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.675770 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-vlgwn"] Dec 03 19:02:51 crc kubenswrapper[4758]: E1203 19:02:51.676194 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe9768c-6273-4788-8aa7-8ed927817cc5" containerName="bootstrap-openstack-openstack-cell1" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.676210 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe9768c-6273-4788-8aa7-8ed927817cc5" containerName="bootstrap-openstack-openstack-cell1" Dec 03 19:02:51 crc kubenswrapper[4758]: E1203 19:02:51.676238 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee10a9b-0fcf-4701-802d-480e6131feed" containerName="keystone-cron" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.676245 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee10a9b-0fcf-4701-802d-480e6131feed" containerName="keystone-cron" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.676437 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ee10a9b-0fcf-4701-802d-480e6131feed" containerName="keystone-cron" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.676468 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe9768c-6273-4788-8aa7-8ed927817cc5" containerName="bootstrap-openstack-openstack-cell1" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.678475 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.683915 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.684134 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.691028 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.693445 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.717417 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-vlgwn"] Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.824253 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ssh-key\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.824732 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-inventory\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.824783 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ceph\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.824814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr5q2\" (UniqueName: \"kubernetes.io/projected/8f4a384c-e511-43ab-b0e0-8def30abdcd4-kube-api-access-lr5q2\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.927745 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ssh-key\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.927863 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-inventory\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.927906 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ceph\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.927934 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr5q2\" (UniqueName: \"kubernetes.io/projected/8f4a384c-e511-43ab-b0e0-8def30abdcd4-kube-api-access-lr5q2\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.937249 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ssh-key\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.937794 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-inventory\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.939303 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ceph\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.945765 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr5q2\" (UniqueName: \"kubernetes.io/projected/8f4a384c-e511-43ab-b0e0-8def30abdcd4-kube-api-access-lr5q2\") pod \"download-cache-openstack-openstack-cell1-vlgwn\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:51 crc kubenswrapper[4758]: I1203 19:02:51.999089 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:02:52 crc kubenswrapper[4758]: I1203 19:02:52.542935 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-vlgwn"] Dec 03 19:02:52 crc kubenswrapper[4758]: I1203 19:02:52.592961 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" event={"ID":"8f4a384c-e511-43ab-b0e0-8def30abdcd4","Type":"ContainerStarted","Data":"2977dc2e887c91536f2ac48ecb836d4cabdae8f0a2303bee2f4a7913443e9c4a"} Dec 03 19:02:53 crc kubenswrapper[4758]: I1203 19:02:53.604908 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" event={"ID":"8f4a384c-e511-43ab-b0e0-8def30abdcd4","Type":"ContainerStarted","Data":"1b82c560d98c01bd3d05041c9e29aaeb6f8080c399d679671fe9f5e755924755"} Dec 03 19:02:53 crc kubenswrapper[4758]: I1203 19:02:53.629462 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" podStartSLOduration=2.367708007 podStartE2EDuration="2.629442383s" podCreationTimestamp="2025-12-03 19:02:51 +0000 UTC" firstStartedPulling="2025-12-03 19:02:52.554183337 +0000 UTC m=+7627.755560218" lastFinishedPulling="2025-12-03 19:02:52.815917703 +0000 UTC m=+7628.017294594" observedRunningTime="2025-12-03 19:02:53.622143985 +0000 UTC m=+7628.823520846" watchObservedRunningTime="2025-12-03 19:02:53.629442383 +0000 UTC m=+7628.830819254" Dec 03 19:03:11 crc kubenswrapper[4758]: I1203 19:03:11.394246 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:03:11 crc kubenswrapper[4758]: I1203 19:03:11.394910 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.506425 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kzspn"] Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.512751 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.522951 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzspn"] Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.670614 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-utilities\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.670767 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-catalog-content\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.670834 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf659\" (UniqueName: \"kubernetes.io/projected/d380d7da-e00f-4b28-99b9-a9ada0c965f4-kube-api-access-nf659\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.773024 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-utilities\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.773158 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-catalog-content\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.773189 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf659\" (UniqueName: \"kubernetes.io/projected/d380d7da-e00f-4b28-99b9-a9ada0c965f4-kube-api-access-nf659\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.773645 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-utilities\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.773794 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-catalog-content\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.794869 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf659\" (UniqueName: \"kubernetes.io/projected/d380d7da-e00f-4b28-99b9-a9ada0c965f4-kube-api-access-nf659\") pod \"community-operators-kzspn\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:34 crc kubenswrapper[4758]: I1203 19:03:34.898259 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:35 crc kubenswrapper[4758]: I1203 19:03:35.471919 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzspn"] Dec 03 19:03:36 crc kubenswrapper[4758]: I1203 19:03:36.215984 4758 generic.go:334] "Generic (PLEG): container finished" podID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerID="054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3" exitCode=0 Dec 03 19:03:36 crc kubenswrapper[4758]: I1203 19:03:36.216026 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzspn" event={"ID":"d380d7da-e00f-4b28-99b9-a9ada0c965f4","Type":"ContainerDied","Data":"054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3"} Dec 03 19:03:36 crc kubenswrapper[4758]: I1203 19:03:36.216280 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzspn" event={"ID":"d380d7da-e00f-4b28-99b9-a9ada0c965f4","Type":"ContainerStarted","Data":"d19d40d66c090beb8b803ee0ca0192fa8f94c22495d2e532efc8ecda00d518fd"} Dec 03 19:03:38 crc kubenswrapper[4758]: I1203 19:03:38.241117 4758 generic.go:334] "Generic (PLEG): container finished" podID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerID="619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5" exitCode=0 Dec 03 19:03:38 crc kubenswrapper[4758]: I1203 19:03:38.241222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzspn" event={"ID":"d380d7da-e00f-4b28-99b9-a9ada0c965f4","Type":"ContainerDied","Data":"619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5"} Dec 03 19:03:39 crc kubenswrapper[4758]: I1203 19:03:39.256220 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzspn" event={"ID":"d380d7da-e00f-4b28-99b9-a9ada0c965f4","Type":"ContainerStarted","Data":"5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8"} Dec 03 19:03:39 crc kubenswrapper[4758]: I1203 19:03:39.281562 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kzspn" podStartSLOduration=2.494398181 podStartE2EDuration="5.281543053s" podCreationTimestamp="2025-12-03 19:03:34 +0000 UTC" firstStartedPulling="2025-12-03 19:03:36.218396166 +0000 UTC m=+7671.419773027" lastFinishedPulling="2025-12-03 19:03:39.005541028 +0000 UTC m=+7674.206917899" observedRunningTime="2025-12-03 19:03:39.281146052 +0000 UTC m=+7674.482522913" watchObservedRunningTime="2025-12-03 19:03:39.281543053 +0000 UTC m=+7674.482919914" Dec 03 19:03:41 crc kubenswrapper[4758]: I1203 19:03:41.394752 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:03:41 crc kubenswrapper[4758]: I1203 19:03:41.395823 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:03:44 crc kubenswrapper[4758]: I1203 19:03:44.900010 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:44 crc kubenswrapper[4758]: I1203 19:03:44.901327 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:44 crc kubenswrapper[4758]: I1203 19:03:44.946381 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:45 crc kubenswrapper[4758]: I1203 19:03:45.377786 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:45 crc kubenswrapper[4758]: I1203 19:03:45.432924 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kzspn"] Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.345397 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kzspn" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="registry-server" containerID="cri-o://5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8" gracePeriod=2 Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.852038 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.956115 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf659\" (UniqueName: \"kubernetes.io/projected/d380d7da-e00f-4b28-99b9-a9ada0c965f4-kube-api-access-nf659\") pod \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.956229 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-catalog-content\") pod \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.956442 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-utilities\") pod \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\" (UID: \"d380d7da-e00f-4b28-99b9-a9ada0c965f4\") " Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.957311 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-utilities" (OuterVolumeSpecName: "utilities") pod "d380d7da-e00f-4b28-99b9-a9ada0c965f4" (UID: "d380d7da-e00f-4b28-99b9-a9ada0c965f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:03:47 crc kubenswrapper[4758]: I1203 19:03:47.961861 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d380d7da-e00f-4b28-99b9-a9ada0c965f4-kube-api-access-nf659" (OuterVolumeSpecName: "kube-api-access-nf659") pod "d380d7da-e00f-4b28-99b9-a9ada0c965f4" (UID: "d380d7da-e00f-4b28-99b9-a9ada0c965f4"). InnerVolumeSpecName "kube-api-access-nf659". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.016181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d380d7da-e00f-4b28-99b9-a9ada0c965f4" (UID: "d380d7da-e00f-4b28-99b9-a9ada0c965f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.059857 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf659\" (UniqueName: \"kubernetes.io/projected/d380d7da-e00f-4b28-99b9-a9ada0c965f4-kube-api-access-nf659\") on node \"crc\" DevicePath \"\"" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.059913 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.059932 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d380d7da-e00f-4b28-99b9-a9ada0c965f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.362706 4758 generic.go:334] "Generic (PLEG): container finished" podID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerID="5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8" exitCode=0 Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.362800 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzspn" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.362831 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzspn" event={"ID":"d380d7da-e00f-4b28-99b9-a9ada0c965f4","Type":"ContainerDied","Data":"5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8"} Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.363978 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzspn" event={"ID":"d380d7da-e00f-4b28-99b9-a9ada0c965f4","Type":"ContainerDied","Data":"d19d40d66c090beb8b803ee0ca0192fa8f94c22495d2e532efc8ecda00d518fd"} Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.364012 4758 scope.go:117] "RemoveContainer" containerID="5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.404823 4758 scope.go:117] "RemoveContainer" containerID="619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.410353 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kzspn"] Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.422232 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kzspn"] Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.448280 4758 scope.go:117] "RemoveContainer" containerID="054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.497635 4758 scope.go:117] "RemoveContainer" containerID="5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8" Dec 03 19:03:48 crc kubenswrapper[4758]: E1203 19:03:48.498334 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8\": container with ID starting with 5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8 not found: ID does not exist" containerID="5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.498386 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8"} err="failed to get container status \"5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8\": rpc error: code = NotFound desc = could not find container \"5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8\": container with ID starting with 5a4dce6cd32a143e5a6448890d6cda63b3dc3b8926648cd744faa2353c9c3eb8 not found: ID does not exist" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.498416 4758 scope.go:117] "RemoveContainer" containerID="619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5" Dec 03 19:03:48 crc kubenswrapper[4758]: E1203 19:03:48.498968 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5\": container with ID starting with 619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5 not found: ID does not exist" containerID="619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.499000 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5"} err="failed to get container status \"619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5\": rpc error: code = NotFound desc = could not find container \"619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5\": container with ID starting with 619e91620bf9b8a4496d0a79161dfaf4affaf23039c8a8db35cf89a6bf7304a5 not found: ID does not exist" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.499079 4758 scope.go:117] "RemoveContainer" containerID="054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3" Dec 03 19:03:48 crc kubenswrapper[4758]: E1203 19:03:48.499368 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3\": container with ID starting with 054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3 not found: ID does not exist" containerID="054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3" Dec 03 19:03:48 crc kubenswrapper[4758]: I1203 19:03:48.499395 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3"} err="failed to get container status \"054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3\": rpc error: code = NotFound desc = could not find container \"054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3\": container with ID starting with 054c6de841b94da362665dbd51ff7b1018f109f1530370342b9d23cf69c3e6f3 not found: ID does not exist" Dec 03 19:03:49 crc kubenswrapper[4758]: I1203 19:03:49.128156 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" path="/var/lib/kubelet/pods/d380d7da-e00f-4b28-99b9-a9ada0c965f4/volumes" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.014959 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wjrh7"] Dec 03 19:04:07 crc kubenswrapper[4758]: E1203 19:04:07.015994 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="extract-content" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.016008 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="extract-content" Dec 03 19:04:07 crc kubenswrapper[4758]: E1203 19:04:07.016030 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="registry-server" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.016036 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="registry-server" Dec 03 19:04:07 crc kubenswrapper[4758]: E1203 19:04:07.016055 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="extract-utilities" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.016062 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="extract-utilities" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.016322 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d380d7da-e00f-4b28-99b9-a9ada0c965f4" containerName="registry-server" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.019027 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.085534 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-utilities\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.085585 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9j5k\" (UniqueName: \"kubernetes.io/projected/c5bf0173-8a67-4003-935f-e527348e94a9-kube-api-access-q9j5k\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.085657 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-catalog-content\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.097480 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjrh7"] Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.187453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-catalog-content\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.188242 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-utilities\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.188281 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9j5k\" (UniqueName: \"kubernetes.io/projected/c5bf0173-8a67-4003-935f-e527348e94a9-kube-api-access-q9j5k\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.188424 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-catalog-content\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.188761 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-utilities\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.209670 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9j5k\" (UniqueName: \"kubernetes.io/projected/c5bf0173-8a67-4003-935f-e527348e94a9-kube-api-access-q9j5k\") pod \"redhat-operators-wjrh7\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.397766 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:07 crc kubenswrapper[4758]: I1203 19:04:07.872140 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjrh7"] Dec 03 19:04:08 crc kubenswrapper[4758]: I1203 19:04:08.605775 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerStarted","Data":"35a04fa5a26675d2d78eecd3ae925ed2f3b65e8bfffc31d3547abd6e13fe0fe9"} Dec 03 19:04:09 crc kubenswrapper[4758]: I1203 19:04:09.616533 4758 generic.go:334] "Generic (PLEG): container finished" podID="c5bf0173-8a67-4003-935f-e527348e94a9" containerID="b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f" exitCode=0 Dec 03 19:04:09 crc kubenswrapper[4758]: I1203 19:04:09.616646 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerDied","Data":"b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f"} Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.394381 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.395173 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.395240 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.396201 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27aab133a3e8862abe59f22c91cb6fa989efdafc746712ea304043f43c4083a5"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.396313 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://27aab133a3e8862abe59f22c91cb6fa989efdafc746712ea304043f43c4083a5" gracePeriod=600 Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.637765 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerStarted","Data":"ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428"} Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.646478 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="27aab133a3e8862abe59f22c91cb6fa989efdafc746712ea304043f43c4083a5" exitCode=0 Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.646537 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"27aab133a3e8862abe59f22c91cb6fa989efdafc746712ea304043f43c4083a5"} Dec 03 19:04:11 crc kubenswrapper[4758]: I1203 19:04:11.646585 4758 scope.go:117] "RemoveContainer" containerID="37cbc88a09f67d7e25bd08e5a3679d6d0cd2d869e5b550df4a4466a4355b0b74" Dec 03 19:04:12 crc kubenswrapper[4758]: I1203 19:04:12.661923 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1"} Dec 03 19:04:14 crc kubenswrapper[4758]: I1203 19:04:14.691399 4758 generic.go:334] "Generic (PLEG): container finished" podID="c5bf0173-8a67-4003-935f-e527348e94a9" containerID="ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428" exitCode=0 Dec 03 19:04:14 crc kubenswrapper[4758]: I1203 19:04:14.691518 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerDied","Data":"ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428"} Dec 03 19:04:15 crc kubenswrapper[4758]: I1203 19:04:15.704270 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerStarted","Data":"e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3"} Dec 03 19:04:16 crc kubenswrapper[4758]: I1203 19:04:16.740784 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wjrh7" podStartSLOduration=5.039996555 podStartE2EDuration="10.740766829s" podCreationTimestamp="2025-12-03 19:04:06 +0000 UTC" firstStartedPulling="2025-12-03 19:04:09.618630718 +0000 UTC m=+7704.820007609" lastFinishedPulling="2025-12-03 19:04:15.319401022 +0000 UTC m=+7710.520777883" observedRunningTime="2025-12-03 19:04:16.735759864 +0000 UTC m=+7711.937136785" watchObservedRunningTime="2025-12-03 19:04:16.740766829 +0000 UTC m=+7711.942143690" Dec 03 19:04:17 crc kubenswrapper[4758]: I1203 19:04:17.398191 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:17 crc kubenswrapper[4758]: I1203 19:04:17.398583 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:18 crc kubenswrapper[4758]: I1203 19:04:18.457164 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wjrh7" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="registry-server" probeResult="failure" output=< Dec 03 19:04:18 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 19:04:18 crc kubenswrapper[4758]: > Dec 03 19:04:23 crc kubenswrapper[4758]: I1203 19:04:23.791161 4758 generic.go:334] "Generic (PLEG): container finished" podID="8f4a384c-e511-43ab-b0e0-8def30abdcd4" containerID="1b82c560d98c01bd3d05041c9e29aaeb6f8080c399d679671fe9f5e755924755" exitCode=0 Dec 03 19:04:23 crc kubenswrapper[4758]: I1203 19:04:23.791270 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" event={"ID":"8f4a384c-e511-43ab-b0e0-8def30abdcd4","Type":"ContainerDied","Data":"1b82c560d98c01bd3d05041c9e29aaeb6f8080c399d679671fe9f5e755924755"} Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.273299 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.371814 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ssh-key\") pod \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.371951 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr5q2\" (UniqueName: \"kubernetes.io/projected/8f4a384c-e511-43ab-b0e0-8def30abdcd4-kube-api-access-lr5q2\") pod \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.372096 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-inventory\") pod \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.372198 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ceph\") pod \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\" (UID: \"8f4a384c-e511-43ab-b0e0-8def30abdcd4\") " Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.377006 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f4a384c-e511-43ab-b0e0-8def30abdcd4-kube-api-access-lr5q2" (OuterVolumeSpecName: "kube-api-access-lr5q2") pod "8f4a384c-e511-43ab-b0e0-8def30abdcd4" (UID: "8f4a384c-e511-43ab-b0e0-8def30abdcd4"). InnerVolumeSpecName "kube-api-access-lr5q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.379053 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ceph" (OuterVolumeSpecName: "ceph") pod "8f4a384c-e511-43ab-b0e0-8def30abdcd4" (UID: "8f4a384c-e511-43ab-b0e0-8def30abdcd4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.403936 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f4a384c-e511-43ab-b0e0-8def30abdcd4" (UID: "8f4a384c-e511-43ab-b0e0-8def30abdcd4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.432845 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-inventory" (OuterVolumeSpecName: "inventory") pod "8f4a384c-e511-43ab-b0e0-8def30abdcd4" (UID: "8f4a384c-e511-43ab-b0e0-8def30abdcd4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.475051 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr5q2\" (UniqueName: \"kubernetes.io/projected/8f4a384c-e511-43ab-b0e0-8def30abdcd4-kube-api-access-lr5q2\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.475085 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.475094 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.475103 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f4a384c-e511-43ab-b0e0-8def30abdcd4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.814374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" event={"ID":"8f4a384c-e511-43ab-b0e0-8def30abdcd4","Type":"ContainerDied","Data":"2977dc2e887c91536f2ac48ecb836d4cabdae8f0a2303bee2f4a7913443e9c4a"} Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.814414 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2977dc2e887c91536f2ac48ecb836d4cabdae8f0a2303bee2f4a7913443e9c4a" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.814478 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-vlgwn" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.919595 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-82l6j"] Dec 03 19:04:25 crc kubenswrapper[4758]: E1203 19:04:25.920392 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4a384c-e511-43ab-b0e0-8def30abdcd4" containerName="download-cache-openstack-openstack-cell1" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.920410 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4a384c-e511-43ab-b0e0-8def30abdcd4" containerName="download-cache-openstack-openstack-cell1" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.920615 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4a384c-e511-43ab-b0e0-8def30abdcd4" containerName="download-cache-openstack-openstack-cell1" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.921443 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.926423 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.927322 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.927927 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.928541 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:04:25 crc kubenswrapper[4758]: I1203 19:04:25.966407 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-82l6j"] Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.086596 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ceph\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.086992 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj84z\" (UniqueName: \"kubernetes.io/projected/5d46d38d-5c92-409c-bc14-578aca300889-kube-api-access-mj84z\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.087038 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-inventory\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.087075 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ssh-key\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.189193 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj84z\" (UniqueName: \"kubernetes.io/projected/5d46d38d-5c92-409c-bc14-578aca300889-kube-api-access-mj84z\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.189269 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-inventory\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.189313 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ssh-key\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.189495 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ceph\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.194460 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-inventory\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.195024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ssh-key\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.196392 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ceph\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.209219 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj84z\" (UniqueName: \"kubernetes.io/projected/5d46d38d-5c92-409c-bc14-578aca300889-kube-api-access-mj84z\") pod \"configure-network-openstack-openstack-cell1-82l6j\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.292318 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.900253 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-82l6j"] Dec 03 19:04:26 crc kubenswrapper[4758]: I1203 19:04:26.915128 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:04:27 crc kubenswrapper[4758]: I1203 19:04:27.468093 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:27 crc kubenswrapper[4758]: I1203 19:04:27.547962 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:27 crc kubenswrapper[4758]: I1203 19:04:27.721883 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wjrh7"] Dec 03 19:04:27 crc kubenswrapper[4758]: I1203 19:04:27.858865 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" event={"ID":"5d46d38d-5c92-409c-bc14-578aca300889","Type":"ContainerStarted","Data":"900b2f9405dd19a39abcd6547a1e723ddef9de851229154ae7724fb63e3a86b3"} Dec 03 19:04:27 crc kubenswrapper[4758]: I1203 19:04:27.858948 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" event={"ID":"5d46d38d-5c92-409c-bc14-578aca300889","Type":"ContainerStarted","Data":"b4ddc0f6400d150122d4d1d3fbe1d1ce4c2f54444591e5302635ad52b4875b5e"} Dec 03 19:04:27 crc kubenswrapper[4758]: I1203 19:04:27.905502 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" podStartSLOduration=2.71215693 podStartE2EDuration="2.905470487s" podCreationTimestamp="2025-12-03 19:04:25 +0000 UTC" firstStartedPulling="2025-12-03 19:04:26.914865063 +0000 UTC m=+7722.116241924" lastFinishedPulling="2025-12-03 19:04:27.10817862 +0000 UTC m=+7722.309555481" observedRunningTime="2025-12-03 19:04:27.884758764 +0000 UTC m=+7723.086135675" watchObservedRunningTime="2025-12-03 19:04:27.905470487 +0000 UTC m=+7723.106847378" Dec 03 19:04:28 crc kubenswrapper[4758]: I1203 19:04:28.870610 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wjrh7" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="registry-server" containerID="cri-o://e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3" gracePeriod=2 Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.447807 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.577811 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-catalog-content\") pod \"c5bf0173-8a67-4003-935f-e527348e94a9\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.578289 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9j5k\" (UniqueName: \"kubernetes.io/projected/c5bf0173-8a67-4003-935f-e527348e94a9-kube-api-access-q9j5k\") pod \"c5bf0173-8a67-4003-935f-e527348e94a9\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.578414 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-utilities\") pod \"c5bf0173-8a67-4003-935f-e527348e94a9\" (UID: \"c5bf0173-8a67-4003-935f-e527348e94a9\") " Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.579450 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-utilities" (OuterVolumeSpecName: "utilities") pod "c5bf0173-8a67-4003-935f-e527348e94a9" (UID: "c5bf0173-8a67-4003-935f-e527348e94a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.584897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5bf0173-8a67-4003-935f-e527348e94a9-kube-api-access-q9j5k" (OuterVolumeSpecName: "kube-api-access-q9j5k") pod "c5bf0173-8a67-4003-935f-e527348e94a9" (UID: "c5bf0173-8a67-4003-935f-e527348e94a9"). InnerVolumeSpecName "kube-api-access-q9j5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.681435 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9j5k\" (UniqueName: \"kubernetes.io/projected/c5bf0173-8a67-4003-935f-e527348e94a9-kube-api-access-q9j5k\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.681471 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.707986 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5bf0173-8a67-4003-935f-e527348e94a9" (UID: "c5bf0173-8a67-4003-935f-e527348e94a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.783192 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bf0173-8a67-4003-935f-e527348e94a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.887010 4758 generic.go:334] "Generic (PLEG): container finished" podID="c5bf0173-8a67-4003-935f-e527348e94a9" containerID="e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3" exitCode=0 Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.887086 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjrh7" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.887084 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerDied","Data":"e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3"} Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.887314 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjrh7" event={"ID":"c5bf0173-8a67-4003-935f-e527348e94a9","Type":"ContainerDied","Data":"35a04fa5a26675d2d78eecd3ae925ed2f3b65e8bfffc31d3547abd6e13fe0fe9"} Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.887337 4758 scope.go:117] "RemoveContainer" containerID="e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.920784 4758 scope.go:117] "RemoveContainer" containerID="ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428" Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.947905 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wjrh7"] Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.957652 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wjrh7"] Dec 03 19:04:29 crc kubenswrapper[4758]: I1203 19:04:29.978167 4758 scope.go:117] "RemoveContainer" containerID="b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f" Dec 03 19:04:30 crc kubenswrapper[4758]: I1203 19:04:30.039567 4758 scope.go:117] "RemoveContainer" containerID="e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3" Dec 03 19:04:30 crc kubenswrapper[4758]: E1203 19:04:30.040615 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3\": container with ID starting with e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3 not found: ID does not exist" containerID="e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3" Dec 03 19:04:30 crc kubenswrapper[4758]: I1203 19:04:30.040672 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3"} err="failed to get container status \"e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3\": rpc error: code = NotFound desc = could not find container \"e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3\": container with ID starting with e91d7f777a33486c67c7be748c8104e1fd15e4fe51c0db188f8cc489514b0ed3 not found: ID does not exist" Dec 03 19:04:30 crc kubenswrapper[4758]: I1203 19:04:30.040723 4758 scope.go:117] "RemoveContainer" containerID="ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428" Dec 03 19:04:30 crc kubenswrapper[4758]: E1203 19:04:30.044841 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428\": container with ID starting with ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428 not found: ID does not exist" containerID="ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428" Dec 03 19:04:30 crc kubenswrapper[4758]: I1203 19:04:30.044872 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428"} err="failed to get container status \"ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428\": rpc error: code = NotFound desc = could not find container \"ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428\": container with ID starting with ada44cc3b16fae1a104976fcfca9a0db7cc697506412b924fde7edc24652d428 not found: ID does not exist" Dec 03 19:04:30 crc kubenswrapper[4758]: I1203 19:04:30.044897 4758 scope.go:117] "RemoveContainer" containerID="b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f" Dec 03 19:04:30 crc kubenswrapper[4758]: E1203 19:04:30.045489 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f\": container with ID starting with b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f not found: ID does not exist" containerID="b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f" Dec 03 19:04:30 crc kubenswrapper[4758]: I1203 19:04:30.045574 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f"} err="failed to get container status \"b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f\": rpc error: code = NotFound desc = could not find container \"b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f\": container with ID starting with b51839a3a9aa2d305aa3ea12202caa5cb1e629495dafd1ca567c9b91f4d9fe8f not found: ID does not exist" Dec 03 19:04:31 crc kubenswrapper[4758]: I1203 19:04:31.138523 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" path="/var/lib/kubelet/pods/c5bf0173-8a67-4003-935f-e527348e94a9/volumes" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.167416 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p8xhr"] Dec 03 19:04:33 crc kubenswrapper[4758]: E1203 19:04:33.170755 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="extract-content" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.170801 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="extract-content" Dec 03 19:04:33 crc kubenswrapper[4758]: E1203 19:04:33.170950 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="registry-server" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.170978 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="registry-server" Dec 03 19:04:33 crc kubenswrapper[4758]: E1203 19:04:33.171017 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="extract-utilities" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.171031 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="extract-utilities" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.171854 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bf0173-8a67-4003-935f-e527348e94a9" containerName="registry-server" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.177779 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.191145 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8xhr"] Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.286585 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-utilities\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.286656 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-catalog-content\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.286734 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lbbv\" (UniqueName: \"kubernetes.io/projected/029d7cd3-30ab-4407-81d7-fcc23cf8e853-kube-api-access-7lbbv\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.389213 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lbbv\" (UniqueName: \"kubernetes.io/projected/029d7cd3-30ab-4407-81d7-fcc23cf8e853-kube-api-access-7lbbv\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.389482 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-utilities\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.389536 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-catalog-content\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.390182 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-catalog-content\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.390197 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-utilities\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.417833 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lbbv\" (UniqueName: \"kubernetes.io/projected/029d7cd3-30ab-4407-81d7-fcc23cf8e853-kube-api-access-7lbbv\") pod \"redhat-marketplace-p8xhr\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:33 crc kubenswrapper[4758]: I1203 19:04:33.513563 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:34 crc kubenswrapper[4758]: I1203 19:04:34.052103 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8xhr"] Dec 03 19:04:34 crc kubenswrapper[4758]: I1203 19:04:34.949806 4758 generic.go:334] "Generic (PLEG): container finished" podID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerID="17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866" exitCode=0 Dec 03 19:04:34 crc kubenswrapper[4758]: I1203 19:04:34.949888 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerDied","Data":"17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866"} Dec 03 19:04:34 crc kubenswrapper[4758]: I1203 19:04:34.950129 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerStarted","Data":"1da6d950b95fd67e6ba410657cb9c52c3f83410913c601dda5ee99c783450bf0"} Dec 03 19:04:35 crc kubenswrapper[4758]: I1203 19:04:35.963562 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerStarted","Data":"d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0"} Dec 03 19:04:36 crc kubenswrapper[4758]: I1203 19:04:36.975533 4758 generic.go:334] "Generic (PLEG): container finished" podID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerID="d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0" exitCode=0 Dec 03 19:04:36 crc kubenswrapper[4758]: I1203 19:04:36.975675 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerDied","Data":"d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0"} Dec 03 19:04:37 crc kubenswrapper[4758]: I1203 19:04:37.997358 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerStarted","Data":"67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5"} Dec 03 19:04:38 crc kubenswrapper[4758]: I1203 19:04:38.029387 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p8xhr" podStartSLOduration=2.615712858 podStartE2EDuration="5.029364396s" podCreationTimestamp="2025-12-03 19:04:33 +0000 UTC" firstStartedPulling="2025-12-03 19:04:34.954182661 +0000 UTC m=+7730.155559522" lastFinishedPulling="2025-12-03 19:04:37.367834199 +0000 UTC m=+7732.569211060" observedRunningTime="2025-12-03 19:04:38.021267396 +0000 UTC m=+7733.222644257" watchObservedRunningTime="2025-12-03 19:04:38.029364396 +0000 UTC m=+7733.230741277" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.340317 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bhspv"] Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.349015 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.382129 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhspv"] Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.478013 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkzpk\" (UniqueName: \"kubernetes.io/projected/d78475c2-e365-44f0-beee-7bb482fdc169-kube-api-access-pkzpk\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.478094 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-utilities\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.478300 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-catalog-content\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.581309 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-catalog-content\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.581480 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkzpk\" (UniqueName: \"kubernetes.io/projected/d78475c2-e365-44f0-beee-7bb482fdc169-kube-api-access-pkzpk\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.581581 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-utilities\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.581938 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-catalog-content\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.581995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-utilities\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.607352 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkzpk\" (UniqueName: \"kubernetes.io/projected/d78475c2-e365-44f0-beee-7bb482fdc169-kube-api-access-pkzpk\") pod \"certified-operators-bhspv\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:41 crc kubenswrapper[4758]: I1203 19:04:41.692736 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:42 crc kubenswrapper[4758]: I1203 19:04:42.292532 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhspv"] Dec 03 19:04:42 crc kubenswrapper[4758]: W1203 19:04:42.296777 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd78475c2_e365_44f0_beee_7bb482fdc169.slice/crio-b21a56c60216751d35d4d2b26d8ab8067d6ae71f036bc6db29e901c2f5ee20c1 WatchSource:0}: Error finding container b21a56c60216751d35d4d2b26d8ab8067d6ae71f036bc6db29e901c2f5ee20c1: Status 404 returned error can't find the container with id b21a56c60216751d35d4d2b26d8ab8067d6ae71f036bc6db29e901c2f5ee20c1 Dec 03 19:04:43 crc kubenswrapper[4758]: I1203 19:04:43.085058 4758 generic.go:334] "Generic (PLEG): container finished" podID="d78475c2-e365-44f0-beee-7bb482fdc169" containerID="ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359" exitCode=0 Dec 03 19:04:43 crc kubenswrapper[4758]: I1203 19:04:43.085150 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerDied","Data":"ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359"} Dec 03 19:04:43 crc kubenswrapper[4758]: I1203 19:04:43.085448 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerStarted","Data":"b21a56c60216751d35d4d2b26d8ab8067d6ae71f036bc6db29e901c2f5ee20c1"} Dec 03 19:04:43 crc kubenswrapper[4758]: I1203 19:04:43.514034 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:43 crc kubenswrapper[4758]: I1203 19:04:43.514479 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:43 crc kubenswrapper[4758]: I1203 19:04:43.605172 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:44 crc kubenswrapper[4758]: I1203 19:04:44.097614 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerStarted","Data":"46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b"} Dec 03 19:04:44 crc kubenswrapper[4758]: I1203 19:04:44.158358 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:45 crc kubenswrapper[4758]: I1203 19:04:45.109419 4758 generic.go:334] "Generic (PLEG): container finished" podID="d78475c2-e365-44f0-beee-7bb482fdc169" containerID="46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b" exitCode=0 Dec 03 19:04:45 crc kubenswrapper[4758]: I1203 19:04:45.109518 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerDied","Data":"46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b"} Dec 03 19:04:45 crc kubenswrapper[4758]: I1203 19:04:45.912404 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8xhr"] Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.121363 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerStarted","Data":"a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547"} Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.121498 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p8xhr" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="registry-server" containerID="cri-o://67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5" gracePeriod=2 Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.147765 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bhspv" podStartSLOduration=2.708361537 podStartE2EDuration="5.147746914s" podCreationTimestamp="2025-12-03 19:04:41 +0000 UTC" firstStartedPulling="2025-12-03 19:04:43.08799501 +0000 UTC m=+7738.289371881" lastFinishedPulling="2025-12-03 19:04:45.527380387 +0000 UTC m=+7740.728757258" observedRunningTime="2025-12-03 19:04:46.14135021 +0000 UTC m=+7741.342727071" watchObservedRunningTime="2025-12-03 19:04:46.147746914 +0000 UTC m=+7741.349123775" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.600770 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.700342 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-utilities\") pod \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.700472 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lbbv\" (UniqueName: \"kubernetes.io/projected/029d7cd3-30ab-4407-81d7-fcc23cf8e853-kube-api-access-7lbbv\") pod \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.700517 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-catalog-content\") pod \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\" (UID: \"029d7cd3-30ab-4407-81d7-fcc23cf8e853\") " Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.701384 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-utilities" (OuterVolumeSpecName: "utilities") pod "029d7cd3-30ab-4407-81d7-fcc23cf8e853" (UID: "029d7cd3-30ab-4407-81d7-fcc23cf8e853"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.706216 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/029d7cd3-30ab-4407-81d7-fcc23cf8e853-kube-api-access-7lbbv" (OuterVolumeSpecName: "kube-api-access-7lbbv") pod "029d7cd3-30ab-4407-81d7-fcc23cf8e853" (UID: "029d7cd3-30ab-4407-81d7-fcc23cf8e853"). InnerVolumeSpecName "kube-api-access-7lbbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.720358 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "029d7cd3-30ab-4407-81d7-fcc23cf8e853" (UID: "029d7cd3-30ab-4407-81d7-fcc23cf8e853"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.803354 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.803747 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lbbv\" (UniqueName: \"kubernetes.io/projected/029d7cd3-30ab-4407-81d7-fcc23cf8e853-kube-api-access-7lbbv\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:46 crc kubenswrapper[4758]: I1203 19:04:46.803764 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029d7cd3-30ab-4407-81d7-fcc23cf8e853-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.135719 4758 generic.go:334] "Generic (PLEG): container finished" podID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerID="67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5" exitCode=0 Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.135799 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerDied","Data":"67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5"} Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.135881 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8xhr" event={"ID":"029d7cd3-30ab-4407-81d7-fcc23cf8e853","Type":"ContainerDied","Data":"1da6d950b95fd67e6ba410657cb9c52c3f83410913c601dda5ee99c783450bf0"} Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.135898 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8xhr" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.135912 4758 scope.go:117] "RemoveContainer" containerID="67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.180702 4758 scope.go:117] "RemoveContainer" containerID="d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.187205 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8xhr"] Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.199339 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8xhr"] Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.207163 4758 scope.go:117] "RemoveContainer" containerID="17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.247504 4758 scope.go:117] "RemoveContainer" containerID="67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5" Dec 03 19:04:47 crc kubenswrapper[4758]: E1203 19:04:47.248126 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5\": container with ID starting with 67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5 not found: ID does not exist" containerID="67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.248194 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5"} err="failed to get container status \"67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5\": rpc error: code = NotFound desc = could not find container \"67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5\": container with ID starting with 67448744802f1549d9f33fae4d739558476c9d23542a75348921da8e4da2bda5 not found: ID does not exist" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.248235 4758 scope.go:117] "RemoveContainer" containerID="d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0" Dec 03 19:04:47 crc kubenswrapper[4758]: E1203 19:04:47.248650 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0\": container with ID starting with d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0 not found: ID does not exist" containerID="d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.248778 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0"} err="failed to get container status \"d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0\": rpc error: code = NotFound desc = could not find container \"d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0\": container with ID starting with d4994ac7fa7568e10aa06cb09f2fc6a0b3536bfa241261edb718bb80ba6f9ff0 not found: ID does not exist" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.248808 4758 scope.go:117] "RemoveContainer" containerID="17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866" Dec 03 19:04:47 crc kubenswrapper[4758]: E1203 19:04:47.249216 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866\": container with ID starting with 17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866 not found: ID does not exist" containerID="17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866" Dec 03 19:04:47 crc kubenswrapper[4758]: I1203 19:04:47.249246 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866"} err="failed to get container status \"17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866\": rpc error: code = NotFound desc = could not find container \"17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866\": container with ID starting with 17e3630b05f1bba1c465431464ec07c22d8e697ce24d3d25636f191c89db8866 not found: ID does not exist" Dec 03 19:04:49 crc kubenswrapper[4758]: I1203 19:04:49.134466 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" path="/var/lib/kubelet/pods/029d7cd3-30ab-4407-81d7-fcc23cf8e853/volumes" Dec 03 19:04:51 crc kubenswrapper[4758]: I1203 19:04:51.693621 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:51 crc kubenswrapper[4758]: I1203 19:04:51.694888 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:51 crc kubenswrapper[4758]: I1203 19:04:51.762983 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:52 crc kubenswrapper[4758]: I1203 19:04:52.233354 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:52 crc kubenswrapper[4758]: I1203 19:04:52.283613 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bhspv"] Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.214893 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bhspv" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="registry-server" containerID="cri-o://a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547" gracePeriod=2 Dec 03 19:04:54 crc kubenswrapper[4758]: E1203 19:04:54.550096 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd78475c2_e365_44f0_beee_7bb482fdc169.slice/crio-a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd78475c2_e365_44f0_beee_7bb482fdc169.slice/crio-conmon-a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547.scope\": RecentStats: unable to find data in memory cache]" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.723363 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.889732 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkzpk\" (UniqueName: \"kubernetes.io/projected/d78475c2-e365-44f0-beee-7bb482fdc169-kube-api-access-pkzpk\") pod \"d78475c2-e365-44f0-beee-7bb482fdc169\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.889862 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-utilities\") pod \"d78475c2-e365-44f0-beee-7bb482fdc169\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.890107 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-catalog-content\") pod \"d78475c2-e365-44f0-beee-7bb482fdc169\" (UID: \"d78475c2-e365-44f0-beee-7bb482fdc169\") " Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.890715 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-utilities" (OuterVolumeSpecName: "utilities") pod "d78475c2-e365-44f0-beee-7bb482fdc169" (UID: "d78475c2-e365-44f0-beee-7bb482fdc169"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.895438 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d78475c2-e365-44f0-beee-7bb482fdc169-kube-api-access-pkzpk" (OuterVolumeSpecName: "kube-api-access-pkzpk") pod "d78475c2-e365-44f0-beee-7bb482fdc169" (UID: "d78475c2-e365-44f0-beee-7bb482fdc169"). InnerVolumeSpecName "kube-api-access-pkzpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.968153 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d78475c2-e365-44f0-beee-7bb482fdc169" (UID: "d78475c2-e365-44f0-beee-7bb482fdc169"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.992264 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.992331 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkzpk\" (UniqueName: \"kubernetes.io/projected/d78475c2-e365-44f0-beee-7bb482fdc169-kube-api-access-pkzpk\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:54 crc kubenswrapper[4758]: I1203 19:04:54.992343 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d78475c2-e365-44f0-beee-7bb482fdc169-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.225938 4758 generic.go:334] "Generic (PLEG): container finished" podID="d78475c2-e365-44f0-beee-7bb482fdc169" containerID="a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547" exitCode=0 Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.226039 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhspv" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.226051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerDied","Data":"a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547"} Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.226330 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhspv" event={"ID":"d78475c2-e365-44f0-beee-7bb482fdc169","Type":"ContainerDied","Data":"b21a56c60216751d35d4d2b26d8ab8067d6ae71f036bc6db29e901c2f5ee20c1"} Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.226353 4758 scope.go:117] "RemoveContainer" containerID="a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.248293 4758 scope.go:117] "RemoveContainer" containerID="46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.256623 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bhspv"] Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.266097 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bhspv"] Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.278998 4758 scope.go:117] "RemoveContainer" containerID="ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.345472 4758 scope.go:117] "RemoveContainer" containerID="a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547" Dec 03 19:04:55 crc kubenswrapper[4758]: E1203 19:04:55.345952 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547\": container with ID starting with a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547 not found: ID does not exist" containerID="a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.345981 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547"} err="failed to get container status \"a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547\": rpc error: code = NotFound desc = could not find container \"a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547\": container with ID starting with a18911335c870dd389e75f6b6953bda618c33c30153a911bcfbc65305e2fe547 not found: ID does not exist" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.346032 4758 scope.go:117] "RemoveContainer" containerID="46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b" Dec 03 19:04:55 crc kubenswrapper[4758]: E1203 19:04:55.346294 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b\": container with ID starting with 46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b not found: ID does not exist" containerID="46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.346341 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b"} err="failed to get container status \"46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b\": rpc error: code = NotFound desc = could not find container \"46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b\": container with ID starting with 46231d10653885b684805a9767e4f57a81ff3ded753835f5e5617616380b0f0b not found: ID does not exist" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.346376 4758 scope.go:117] "RemoveContainer" containerID="ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359" Dec 03 19:04:55 crc kubenswrapper[4758]: E1203 19:04:55.346658 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359\": container with ID starting with ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359 not found: ID does not exist" containerID="ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359" Dec 03 19:04:55 crc kubenswrapper[4758]: I1203 19:04:55.346697 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359"} err="failed to get container status \"ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359\": rpc error: code = NotFound desc = could not find container \"ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359\": container with ID starting with ba24b33bdcea5d5c2644cb926d3d5a62844de547f1bb91ccb501ee217cb4e359 not found: ID does not exist" Dec 03 19:04:57 crc kubenswrapper[4758]: I1203 19:04:57.128264 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" path="/var/lib/kubelet/pods/d78475c2-e365-44f0-beee-7bb482fdc169/volumes" Dec 03 19:05:50 crc kubenswrapper[4758]: I1203 19:05:50.483449 4758 generic.go:334] "Generic (PLEG): container finished" podID="5d46d38d-5c92-409c-bc14-578aca300889" containerID="900b2f9405dd19a39abcd6547a1e723ddef9de851229154ae7724fb63e3a86b3" exitCode=0 Dec 03 19:05:50 crc kubenswrapper[4758]: I1203 19:05:50.483589 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" event={"ID":"5d46d38d-5c92-409c-bc14-578aca300889","Type":"ContainerDied","Data":"900b2f9405dd19a39abcd6547a1e723ddef9de851229154ae7724fb63e3a86b3"} Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.002468 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.164857 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj84z\" (UniqueName: \"kubernetes.io/projected/5d46d38d-5c92-409c-bc14-578aca300889-kube-api-access-mj84z\") pod \"5d46d38d-5c92-409c-bc14-578aca300889\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.164951 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ssh-key\") pod \"5d46d38d-5c92-409c-bc14-578aca300889\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.164975 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-inventory\") pod \"5d46d38d-5c92-409c-bc14-578aca300889\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.165225 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ceph\") pod \"5d46d38d-5c92-409c-bc14-578aca300889\" (UID: \"5d46d38d-5c92-409c-bc14-578aca300889\") " Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.173395 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d46d38d-5c92-409c-bc14-578aca300889-kube-api-access-mj84z" (OuterVolumeSpecName: "kube-api-access-mj84z") pod "5d46d38d-5c92-409c-bc14-578aca300889" (UID: "5d46d38d-5c92-409c-bc14-578aca300889"). InnerVolumeSpecName "kube-api-access-mj84z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.175866 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ceph" (OuterVolumeSpecName: "ceph") pod "5d46d38d-5c92-409c-bc14-578aca300889" (UID: "5d46d38d-5c92-409c-bc14-578aca300889"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.216938 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d46d38d-5c92-409c-bc14-578aca300889" (UID: "5d46d38d-5c92-409c-bc14-578aca300889"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.218879 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-inventory" (OuterVolumeSpecName: "inventory") pod "5d46d38d-5c92-409c-bc14-578aca300889" (UID: "5d46d38d-5c92-409c-bc14-578aca300889"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.267724 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.267754 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj84z\" (UniqueName: \"kubernetes.io/projected/5d46d38d-5c92-409c-bc14-578aca300889-kube-api-access-mj84z\") on node \"crc\" DevicePath \"\"" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.267767 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.267776 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d46d38d-5c92-409c-bc14-578aca300889-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.511419 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" event={"ID":"5d46d38d-5c92-409c-bc14-578aca300889","Type":"ContainerDied","Data":"b4ddc0f6400d150122d4d1d3fbe1d1ce4c2f54444591e5302635ad52b4875b5e"} Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.511837 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4ddc0f6400d150122d4d1d3fbe1d1ce4c2f54444591e5302635ad52b4875b5e" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.511463 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-82l6j" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.600340 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-x6gff"] Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601514 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d46d38d-5c92-409c-bc14-578aca300889" containerName="configure-network-openstack-openstack-cell1" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601572 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d46d38d-5c92-409c-bc14-578aca300889" containerName="configure-network-openstack-openstack-cell1" Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601626 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="extract-utilities" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601645 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="extract-utilities" Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601741 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="extract-content" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601764 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="extract-content" Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601804 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="extract-utilities" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601821 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="extract-utilities" Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601846 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="registry-server" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601868 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="registry-server" Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601929 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="registry-server" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601949 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="registry-server" Dec 03 19:05:52 crc kubenswrapper[4758]: E1203 19:05:52.601978 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="extract-content" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.601995 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="extract-content" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.602484 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d78475c2-e365-44f0-beee-7bb482fdc169" containerName="registry-server" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.602539 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d46d38d-5c92-409c-bc14-578aca300889" containerName="configure-network-openstack-openstack-cell1" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.602577 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="029d7cd3-30ab-4407-81d7-fcc23cf8e853" containerName="registry-server" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.610955 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.619185 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.620419 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.620613 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.621043 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.638256 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-x6gff"] Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.778757 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ceph\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.778842 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-inventory\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.778942 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7bct\" (UniqueName: \"kubernetes.io/projected/57d874e8-5585-46a0-973c-3cfa587a7e3d-kube-api-access-v7bct\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.778977 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ssh-key\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.881633 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7bct\" (UniqueName: \"kubernetes.io/projected/57d874e8-5585-46a0-973c-3cfa587a7e3d-kube-api-access-v7bct\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.882107 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ssh-key\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.882352 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ceph\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.882444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-inventory\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.888707 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-inventory\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.890336 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ssh-key\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.890752 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ceph\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.905775 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7bct\" (UniqueName: \"kubernetes.io/projected/57d874e8-5585-46a0-973c-3cfa587a7e3d-kube-api-access-v7bct\") pod \"validate-network-openstack-openstack-cell1-x6gff\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:52 crc kubenswrapper[4758]: I1203 19:05:52.949237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:05:53 crc kubenswrapper[4758]: I1203 19:05:53.343209 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-x6gff"] Dec 03 19:05:53 crc kubenswrapper[4758]: I1203 19:05:53.542552 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" event={"ID":"57d874e8-5585-46a0-973c-3cfa587a7e3d","Type":"ContainerStarted","Data":"6aec96e9dc99d06ca637b16cb1be3312c9dcf6f327078ad99a18d732ba9b3517"} Dec 03 19:05:54 crc kubenswrapper[4758]: I1203 19:05:54.556160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" event={"ID":"57d874e8-5585-46a0-973c-3cfa587a7e3d","Type":"ContainerStarted","Data":"b643c55914d3fe1af9e751da8b823f9ba1394508a8d512c1be699e0d7adc1449"} Dec 03 19:05:54 crc kubenswrapper[4758]: I1203 19:05:54.577058 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" podStartSLOduration=2.400444745 podStartE2EDuration="2.577038126s" podCreationTimestamp="2025-12-03 19:05:52 +0000 UTC" firstStartedPulling="2025-12-03 19:05:53.352365657 +0000 UTC m=+7808.553742518" lastFinishedPulling="2025-12-03 19:05:53.528959028 +0000 UTC m=+7808.730335899" observedRunningTime="2025-12-03 19:05:54.572812451 +0000 UTC m=+7809.774189312" watchObservedRunningTime="2025-12-03 19:05:54.577038126 +0000 UTC m=+7809.778414987" Dec 03 19:05:58 crc kubenswrapper[4758]: I1203 19:05:58.597952 4758 generic.go:334] "Generic (PLEG): container finished" podID="57d874e8-5585-46a0-973c-3cfa587a7e3d" containerID="b643c55914d3fe1af9e751da8b823f9ba1394508a8d512c1be699e0d7adc1449" exitCode=0 Dec 03 19:05:58 crc kubenswrapper[4758]: I1203 19:05:58.598019 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" event={"ID":"57d874e8-5585-46a0-973c-3cfa587a7e3d","Type":"ContainerDied","Data":"b643c55914d3fe1af9e751da8b823f9ba1394508a8d512c1be699e0d7adc1449"} Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.053006 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.165051 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ssh-key\") pod \"57d874e8-5585-46a0-973c-3cfa587a7e3d\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.165154 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-inventory\") pod \"57d874e8-5585-46a0-973c-3cfa587a7e3d\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.165191 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7bct\" (UniqueName: \"kubernetes.io/projected/57d874e8-5585-46a0-973c-3cfa587a7e3d-kube-api-access-v7bct\") pod \"57d874e8-5585-46a0-973c-3cfa587a7e3d\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.165244 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ceph\") pod \"57d874e8-5585-46a0-973c-3cfa587a7e3d\" (UID: \"57d874e8-5585-46a0-973c-3cfa587a7e3d\") " Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.171134 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d874e8-5585-46a0-973c-3cfa587a7e3d-kube-api-access-v7bct" (OuterVolumeSpecName: "kube-api-access-v7bct") pod "57d874e8-5585-46a0-973c-3cfa587a7e3d" (UID: "57d874e8-5585-46a0-973c-3cfa587a7e3d"). InnerVolumeSpecName "kube-api-access-v7bct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.176798 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ceph" (OuterVolumeSpecName: "ceph") pod "57d874e8-5585-46a0-973c-3cfa587a7e3d" (UID: "57d874e8-5585-46a0-973c-3cfa587a7e3d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.194254 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57d874e8-5585-46a0-973c-3cfa587a7e3d" (UID: "57d874e8-5585-46a0-973c-3cfa587a7e3d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.194736 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-inventory" (OuterVolumeSpecName: "inventory") pod "57d874e8-5585-46a0-973c-3cfa587a7e3d" (UID: "57d874e8-5585-46a0-973c-3cfa587a7e3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.268956 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.269004 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7bct\" (UniqueName: \"kubernetes.io/projected/57d874e8-5585-46a0-973c-3cfa587a7e3d-kube-api-access-v7bct\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.269024 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.269040 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d874e8-5585-46a0-973c-3cfa587a7e3d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.620531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" event={"ID":"57d874e8-5585-46a0-973c-3cfa587a7e3d","Type":"ContainerDied","Data":"6aec96e9dc99d06ca637b16cb1be3312c9dcf6f327078ad99a18d732ba9b3517"} Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.620594 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6aec96e9dc99d06ca637b16cb1be3312c9dcf6f327078ad99a18d732ba9b3517" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.621087 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-x6gff" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.724239 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-88rh2"] Dec 03 19:06:00 crc kubenswrapper[4758]: E1203 19:06:00.724769 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d874e8-5585-46a0-973c-3cfa587a7e3d" containerName="validate-network-openstack-openstack-cell1" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.724789 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d874e8-5585-46a0-973c-3cfa587a7e3d" containerName="validate-network-openstack-openstack-cell1" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.725095 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d874e8-5585-46a0-973c-3cfa587a7e3d" containerName="validate-network-openstack-openstack-cell1" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.725990 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.728987 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.729277 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.729506 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.730969 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.759840 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-88rh2"] Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.782012 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ssh-key\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.782076 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-inventory\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.782339 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ceph\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.782581 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s22t\" (UniqueName: \"kubernetes.io/projected/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-kube-api-access-7s22t\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.884362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ssh-key\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.884820 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-inventory\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.884974 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ceph\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.885136 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s22t\" (UniqueName: \"kubernetes.io/projected/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-kube-api-access-7s22t\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.888418 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ssh-key\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.890003 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-inventory\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.890196 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ceph\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:00 crc kubenswrapper[4758]: I1203 19:06:00.903753 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s22t\" (UniqueName: \"kubernetes.io/projected/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-kube-api-access-7s22t\") pod \"install-os-openstack-openstack-cell1-88rh2\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:01 crc kubenswrapper[4758]: I1203 19:06:01.062128 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:01 crc kubenswrapper[4758]: I1203 19:06:01.607918 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-88rh2"] Dec 03 19:06:01 crc kubenswrapper[4758]: I1203 19:06:01.633291 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-88rh2" event={"ID":"ca4d63dd-6c41-43c0-9b30-fcece1c986b5","Type":"ContainerStarted","Data":"c4c814e096f18fdd6ee210154a1d1faad3f7d0d49216e87a8f4407b3f119a595"} Dec 03 19:06:02 crc kubenswrapper[4758]: I1203 19:06:02.650669 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-88rh2" event={"ID":"ca4d63dd-6c41-43c0-9b30-fcece1c986b5","Type":"ContainerStarted","Data":"88f4703ef88f18738d216c176888a74b3d608b40731640771242e389af663e5d"} Dec 03 19:06:02 crc kubenswrapper[4758]: I1203 19:06:02.679434 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-88rh2" podStartSLOduration=2.543750652 podStartE2EDuration="2.67940904s" podCreationTimestamp="2025-12-03 19:06:00 +0000 UTC" firstStartedPulling="2025-12-03 19:06:01.612778068 +0000 UTC m=+7816.814154919" lastFinishedPulling="2025-12-03 19:06:01.748436446 +0000 UTC m=+7816.949813307" observedRunningTime="2025-12-03 19:06:02.670869757 +0000 UTC m=+7817.872246628" watchObservedRunningTime="2025-12-03 19:06:02.67940904 +0000 UTC m=+7817.880785931" Dec 03 19:06:41 crc kubenswrapper[4758]: I1203 19:06:41.397289 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:06:41 crc kubenswrapper[4758]: I1203 19:06:41.397986 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:06:47 crc kubenswrapper[4758]: I1203 19:06:47.179992 4758 generic.go:334] "Generic (PLEG): container finished" podID="ca4d63dd-6c41-43c0-9b30-fcece1c986b5" containerID="88f4703ef88f18738d216c176888a74b3d608b40731640771242e389af663e5d" exitCode=0 Dec 03 19:06:47 crc kubenswrapper[4758]: I1203 19:06:47.180083 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-88rh2" event={"ID":"ca4d63dd-6c41-43c0-9b30-fcece1c986b5","Type":"ContainerDied","Data":"88f4703ef88f18738d216c176888a74b3d608b40731640771242e389af663e5d"} Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.662239 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.848666 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s22t\" (UniqueName: \"kubernetes.io/projected/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-kube-api-access-7s22t\") pod \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.848871 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ssh-key\") pod \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.849052 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-inventory\") pod \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.849126 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ceph\") pod \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\" (UID: \"ca4d63dd-6c41-43c0-9b30-fcece1c986b5\") " Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.857892 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ceph" (OuterVolumeSpecName: "ceph") pod "ca4d63dd-6c41-43c0-9b30-fcece1c986b5" (UID: "ca4d63dd-6c41-43c0-9b30-fcece1c986b5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.859908 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-kube-api-access-7s22t" (OuterVolumeSpecName: "kube-api-access-7s22t") pod "ca4d63dd-6c41-43c0-9b30-fcece1c986b5" (UID: "ca4d63dd-6c41-43c0-9b30-fcece1c986b5"). InnerVolumeSpecName "kube-api-access-7s22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.883645 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-inventory" (OuterVolumeSpecName: "inventory") pod "ca4d63dd-6c41-43c0-9b30-fcece1c986b5" (UID: "ca4d63dd-6c41-43c0-9b30-fcece1c986b5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.894080 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca4d63dd-6c41-43c0-9b30-fcece1c986b5" (UID: "ca4d63dd-6c41-43c0-9b30-fcece1c986b5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.952496 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.952549 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.952566 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s22t\" (UniqueName: \"kubernetes.io/projected/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-kube-api-access-7s22t\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:48 crc kubenswrapper[4758]: I1203 19:06:48.952579 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca4d63dd-6c41-43c0-9b30-fcece1c986b5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.210497 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-88rh2" event={"ID":"ca4d63dd-6c41-43c0-9b30-fcece1c986b5","Type":"ContainerDied","Data":"c4c814e096f18fdd6ee210154a1d1faad3f7d0d49216e87a8f4407b3f119a595"} Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.210810 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4c814e096f18fdd6ee210154a1d1faad3f7d0d49216e87a8f4407b3f119a595" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.210597 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-88rh2" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.308353 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-nlm2b"] Dec 03 19:06:49 crc kubenswrapper[4758]: E1203 19:06:49.311058 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4d63dd-6c41-43c0-9b30-fcece1c986b5" containerName="install-os-openstack-openstack-cell1" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.311078 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4d63dd-6c41-43c0-9b30-fcece1c986b5" containerName="install-os-openstack-openstack-cell1" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.313266 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4d63dd-6c41-43c0-9b30-fcece1c986b5" containerName="install-os-openstack-openstack-cell1" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.316180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.319239 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.320330 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.320616 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.330422 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.362551 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-nlm2b"] Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.464932 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-inventory\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.465361 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ceph\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.465395 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ssh-key\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.465420 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w4kc\" (UniqueName: \"kubernetes.io/projected/aa20be19-9e47-498e-b64f-a4840514eaae-kube-api-access-7w4kc\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.567124 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-inventory\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.567239 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ceph\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.567268 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ssh-key\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.567291 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w4kc\" (UniqueName: \"kubernetes.io/projected/aa20be19-9e47-498e-b64f-a4840514eaae-kube-api-access-7w4kc\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.572283 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ssh-key\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.588564 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ceph\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.588873 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-inventory\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.591767 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w4kc\" (UniqueName: \"kubernetes.io/projected/aa20be19-9e47-498e-b64f-a4840514eaae-kube-api-access-7w4kc\") pod \"configure-os-openstack-openstack-cell1-nlm2b\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:49 crc kubenswrapper[4758]: I1203 19:06:49.643273 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:06:50 crc kubenswrapper[4758]: I1203 19:06:50.257896 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-nlm2b"] Dec 03 19:06:51 crc kubenswrapper[4758]: I1203 19:06:51.233078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" event={"ID":"aa20be19-9e47-498e-b64f-a4840514eaae","Type":"ContainerStarted","Data":"49a0e14ee558125ea6b399bc3c7912b9cdd5768dd367098d1e6453ad0344b742"} Dec 03 19:06:51 crc kubenswrapper[4758]: I1203 19:06:51.233437 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" event={"ID":"aa20be19-9e47-498e-b64f-a4840514eaae","Type":"ContainerStarted","Data":"0bff6f7a8ed92f1dffa105e459ab5b13e24ce18e934f8d8c738c9bfa18136329"} Dec 03 19:06:51 crc kubenswrapper[4758]: I1203 19:06:51.259222 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" podStartSLOduration=1.963504843 podStartE2EDuration="2.259195593s" podCreationTimestamp="2025-12-03 19:06:49 +0000 UTC" firstStartedPulling="2025-12-03 19:06:50.268916927 +0000 UTC m=+7865.470293778" lastFinishedPulling="2025-12-03 19:06:50.564607667 +0000 UTC m=+7865.765984528" observedRunningTime="2025-12-03 19:06:51.250229729 +0000 UTC m=+7866.451606600" watchObservedRunningTime="2025-12-03 19:06:51.259195593 +0000 UTC m=+7866.460572494" Dec 03 19:07:11 crc kubenswrapper[4758]: I1203 19:07:11.394981 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:07:11 crc kubenswrapper[4758]: I1203 19:07:11.395670 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:07:38 crc kubenswrapper[4758]: I1203 19:07:38.838002 4758 generic.go:334] "Generic (PLEG): container finished" podID="aa20be19-9e47-498e-b64f-a4840514eaae" containerID="49a0e14ee558125ea6b399bc3c7912b9cdd5768dd367098d1e6453ad0344b742" exitCode=0 Dec 03 19:07:38 crc kubenswrapper[4758]: I1203 19:07:38.838088 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" event={"ID":"aa20be19-9e47-498e-b64f-a4840514eaae","Type":"ContainerDied","Data":"49a0e14ee558125ea6b399bc3c7912b9cdd5768dd367098d1e6453ad0344b742"} Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.357438 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.484079 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ssh-key\") pod \"aa20be19-9e47-498e-b64f-a4840514eaae\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.484297 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-inventory\") pod \"aa20be19-9e47-498e-b64f-a4840514eaae\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.484426 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ceph\") pod \"aa20be19-9e47-498e-b64f-a4840514eaae\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.484521 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w4kc\" (UniqueName: \"kubernetes.io/projected/aa20be19-9e47-498e-b64f-a4840514eaae-kube-api-access-7w4kc\") pod \"aa20be19-9e47-498e-b64f-a4840514eaae\" (UID: \"aa20be19-9e47-498e-b64f-a4840514eaae\") " Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.489923 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ceph" (OuterVolumeSpecName: "ceph") pod "aa20be19-9e47-498e-b64f-a4840514eaae" (UID: "aa20be19-9e47-498e-b64f-a4840514eaae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.491960 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa20be19-9e47-498e-b64f-a4840514eaae-kube-api-access-7w4kc" (OuterVolumeSpecName: "kube-api-access-7w4kc") pod "aa20be19-9e47-498e-b64f-a4840514eaae" (UID: "aa20be19-9e47-498e-b64f-a4840514eaae"). InnerVolumeSpecName "kube-api-access-7w4kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.514950 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa20be19-9e47-498e-b64f-a4840514eaae" (UID: "aa20be19-9e47-498e-b64f-a4840514eaae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.522733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-inventory" (OuterVolumeSpecName: "inventory") pod "aa20be19-9e47-498e-b64f-a4840514eaae" (UID: "aa20be19-9e47-498e-b64f-a4840514eaae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.587211 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.587252 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.587267 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w4kc\" (UniqueName: \"kubernetes.io/projected/aa20be19-9e47-498e-b64f-a4840514eaae-kube-api-access-7w4kc\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.587281 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa20be19-9e47-498e-b64f-a4840514eaae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.868987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" event={"ID":"aa20be19-9e47-498e-b64f-a4840514eaae","Type":"ContainerDied","Data":"0bff6f7a8ed92f1dffa105e459ab5b13e24ce18e934f8d8c738c9bfa18136329"} Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.869035 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bff6f7a8ed92f1dffa105e459ab5b13e24ce18e934f8d8c738c9bfa18136329" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.869114 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-nlm2b" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.960767 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-ldt7n"] Dec 03 19:07:40 crc kubenswrapper[4758]: E1203 19:07:40.961296 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa20be19-9e47-498e-b64f-a4840514eaae" containerName="configure-os-openstack-openstack-cell1" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.961313 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa20be19-9e47-498e-b64f-a4840514eaae" containerName="configure-os-openstack-openstack-cell1" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.961574 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa20be19-9e47-498e-b64f-a4840514eaae" containerName="configure-os-openstack-openstack-cell1" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.962455 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.965007 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.965196 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.965576 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:07:40 crc kubenswrapper[4758]: I1203 19:07:40.965809 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.025336 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-ldt7n"] Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.102720 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.102788 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-inventory-0\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.102974 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qdtj\" (UniqueName: \"kubernetes.io/projected/cb99c7dc-9a7d-4287-b0f9-9590049db61d-kube-api-access-7qdtj\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.103018 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ceph\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.205268 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qdtj\" (UniqueName: \"kubernetes.io/projected/cb99c7dc-9a7d-4287-b0f9-9590049db61d-kube-api-access-7qdtj\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.205350 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ceph\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.205609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.205655 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-inventory-0\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.210043 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ceph\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.210399 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-inventory-0\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.227370 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.235799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qdtj\" (UniqueName: \"kubernetes.io/projected/cb99c7dc-9a7d-4287-b0f9-9590049db61d-kube-api-access-7qdtj\") pod \"ssh-known-hosts-openstack-ldt7n\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.318430 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.395529 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.395613 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.395720 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.397460 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.397628 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" gracePeriod=600 Dec 03 19:07:41 crc kubenswrapper[4758]: E1203 19:07:41.522700 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.882098 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1"} Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.882160 4758 scope.go:117] "RemoveContainer" containerID="27aab133a3e8862abe59f22c91cb6fa989efdafc746712ea304043f43c4083a5" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.882006 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" exitCode=0 Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.884240 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:07:41 crc kubenswrapper[4758]: E1203 19:07:41.885346 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:07:41 crc kubenswrapper[4758]: I1203 19:07:41.967825 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-ldt7n"] Dec 03 19:07:42 crc kubenswrapper[4758]: I1203 19:07:42.894920 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ldt7n" event={"ID":"cb99c7dc-9a7d-4287-b0f9-9590049db61d","Type":"ContainerStarted","Data":"e64fcf69195d4dcdd88dd1f89c8a45c866fe9f71596c850baf3f6ed3e7f6678a"} Dec 03 19:07:42 crc kubenswrapper[4758]: I1203 19:07:42.895307 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ldt7n" event={"ID":"cb99c7dc-9a7d-4287-b0f9-9590049db61d","Type":"ContainerStarted","Data":"e321de304a107376d028fb0c3595627a06c39834d5b64463451c9380f1c00cd0"} Dec 03 19:07:42 crc kubenswrapper[4758]: I1203 19:07:42.917715 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-ldt7n" podStartSLOduration=2.744113537 podStartE2EDuration="2.917671705s" podCreationTimestamp="2025-12-03 19:07:40 +0000 UTC" firstStartedPulling="2025-12-03 19:07:41.953582352 +0000 UTC m=+7917.154959213" lastFinishedPulling="2025-12-03 19:07:42.12714052 +0000 UTC m=+7917.328517381" observedRunningTime="2025-12-03 19:07:42.908491146 +0000 UTC m=+7918.109868027" watchObservedRunningTime="2025-12-03 19:07:42.917671705 +0000 UTC m=+7918.119048586" Dec 03 19:07:50 crc kubenswrapper[4758]: I1203 19:07:50.989750 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb99c7dc-9a7d-4287-b0f9-9590049db61d" containerID="e64fcf69195d4dcdd88dd1f89c8a45c866fe9f71596c850baf3f6ed3e7f6678a" exitCode=0 Dec 03 19:07:50 crc kubenswrapper[4758]: I1203 19:07:50.989851 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ldt7n" event={"ID":"cb99c7dc-9a7d-4287-b0f9-9590049db61d","Type":"ContainerDied","Data":"e64fcf69195d4dcdd88dd1f89c8a45c866fe9f71596c850baf3f6ed3e7f6678a"} Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.449378 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.576708 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qdtj\" (UniqueName: \"kubernetes.io/projected/cb99c7dc-9a7d-4287-b0f9-9590049db61d-kube-api-access-7qdtj\") pod \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.577216 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-inventory-0\") pod \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.577327 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ceph\") pod \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.577646 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ssh-key-openstack-cell1\") pod \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\" (UID: \"cb99c7dc-9a7d-4287-b0f9-9590049db61d\") " Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.582426 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ceph" (OuterVolumeSpecName: "ceph") pod "cb99c7dc-9a7d-4287-b0f9-9590049db61d" (UID: "cb99c7dc-9a7d-4287-b0f9-9590049db61d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.583936 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb99c7dc-9a7d-4287-b0f9-9590049db61d-kube-api-access-7qdtj" (OuterVolumeSpecName: "kube-api-access-7qdtj") pod "cb99c7dc-9a7d-4287-b0f9-9590049db61d" (UID: "cb99c7dc-9a7d-4287-b0f9-9590049db61d"). InnerVolumeSpecName "kube-api-access-7qdtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.608877 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "cb99c7dc-9a7d-4287-b0f9-9590049db61d" (UID: "cb99c7dc-9a7d-4287-b0f9-9590049db61d"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.618894 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "cb99c7dc-9a7d-4287-b0f9-9590049db61d" (UID: "cb99c7dc-9a7d-4287-b0f9-9590049db61d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.680644 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qdtj\" (UniqueName: \"kubernetes.io/projected/cb99c7dc-9a7d-4287-b0f9-9590049db61d-kube-api-access-7qdtj\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.680913 4758 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.681010 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:52 crc kubenswrapper[4758]: I1203 19:07:52.681104 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb99c7dc-9a7d-4287-b0f9-9590049db61d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.016641 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ldt7n" event={"ID":"cb99c7dc-9a7d-4287-b0f9-9590049db61d","Type":"ContainerDied","Data":"e321de304a107376d028fb0c3595627a06c39834d5b64463451c9380f1c00cd0"} Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.016735 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e321de304a107376d028fb0c3595627a06c39834d5b64463451c9380f1c00cd0" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.016872 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ldt7n" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.109932 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-q9w8s"] Dec 03 19:07:53 crc kubenswrapper[4758]: E1203 19:07:53.110506 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb99c7dc-9a7d-4287-b0f9-9590049db61d" containerName="ssh-known-hosts-openstack" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.110529 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb99c7dc-9a7d-4287-b0f9-9590049db61d" containerName="ssh-known-hosts-openstack" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.111277 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb99c7dc-9a7d-4287-b0f9-9590049db61d" containerName="ssh-known-hosts-openstack" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.112260 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.114648 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.116891 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.117309 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.125230 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.151290 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-q9w8s"] Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.192991 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhbb7\" (UniqueName: \"kubernetes.io/projected/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-kube-api-access-jhbb7\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.193076 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ceph\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.193105 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-inventory\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.193277 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.295747 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.295935 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhbb7\" (UniqueName: \"kubernetes.io/projected/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-kube-api-access-jhbb7\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.295978 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ceph\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.296003 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-inventory\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.300024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.301395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ceph\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.308987 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-inventory\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.318297 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhbb7\" (UniqueName: \"kubernetes.io/projected/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-kube-api-access-jhbb7\") pod \"run-os-openstack-openstack-cell1-q9w8s\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:53 crc kubenswrapper[4758]: I1203 19:07:53.445954 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:07:54 crc kubenswrapper[4758]: I1203 19:07:54.042666 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-q9w8s"] Dec 03 19:07:55 crc kubenswrapper[4758]: I1203 19:07:55.037330 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" event={"ID":"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb","Type":"ContainerStarted","Data":"305b0b7b286f064b92a56bb1c76b9cb1b1624b1610db5f02bf3ad7f6e6c67e04"} Dec 03 19:07:55 crc kubenswrapper[4758]: I1203 19:07:55.037660 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" event={"ID":"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb","Type":"ContainerStarted","Data":"57ec8f7742f7bac69f31a4a1ab85d6d2e2887f56af4c7d88fc98aa2168503e8a"} Dec 03 19:07:55 crc kubenswrapper[4758]: I1203 19:07:55.059317 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" podStartSLOduration=1.86859587 podStartE2EDuration="2.059297186s" podCreationTimestamp="2025-12-03 19:07:53 +0000 UTC" firstStartedPulling="2025-12-03 19:07:54.049143709 +0000 UTC m=+7929.250520610" lastFinishedPulling="2025-12-03 19:07:54.239845055 +0000 UTC m=+7929.441221926" observedRunningTime="2025-12-03 19:07:55.052711406 +0000 UTC m=+7930.254088267" watchObservedRunningTime="2025-12-03 19:07:55.059297186 +0000 UTC m=+7930.260674047" Dec 03 19:07:55 crc kubenswrapper[4758]: I1203 19:07:55.122822 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:07:55 crc kubenswrapper[4758]: E1203 19:07:55.123105 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:08:02 crc kubenswrapper[4758]: I1203 19:08:02.126005 4758 generic.go:334] "Generic (PLEG): container finished" podID="131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" containerID="305b0b7b286f064b92a56bb1c76b9cb1b1624b1610db5f02bf3ad7f6e6c67e04" exitCode=0 Dec 03 19:08:02 crc kubenswrapper[4758]: I1203 19:08:02.126059 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" event={"ID":"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb","Type":"ContainerDied","Data":"305b0b7b286f064b92a56bb1c76b9cb1b1624b1610db5f02bf3ad7f6e6c67e04"} Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.606967 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.744962 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbb7\" (UniqueName: \"kubernetes.io/projected/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-kube-api-access-jhbb7\") pod \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.745265 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ceph\") pod \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.745375 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-inventory\") pod \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.745521 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key\") pod \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.750928 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ceph" (OuterVolumeSpecName: "ceph") pod "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" (UID: "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.752707 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-kube-api-access-jhbb7" (OuterVolumeSpecName: "kube-api-access-jhbb7") pod "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" (UID: "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb"). InnerVolumeSpecName "kube-api-access-jhbb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:08:03 crc kubenswrapper[4758]: E1203 19:08:03.770717 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key podName:131aa418-2634-4ccf-8ae2-5ac72ecfc8eb nodeName:}" failed. No retries permitted until 2025-12-03 19:08:04.270639617 +0000 UTC m=+7939.472016478 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key") pod "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" (UID: "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb") : error deleting /var/lib/kubelet/pods/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb/volume-subpaths: remove /var/lib/kubelet/pods/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb/volume-subpaths: no such file or directory Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.772742 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-inventory" (OuterVolumeSpecName: "inventory") pod "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" (UID: "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.847542 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.847572 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbb7\" (UniqueName: \"kubernetes.io/projected/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-kube-api-access-jhbb7\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:03 crc kubenswrapper[4758]: I1203 19:08:03.847582 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.146179 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" event={"ID":"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb","Type":"ContainerDied","Data":"57ec8f7742f7bac69f31a4a1ab85d6d2e2887f56af4c7d88fc98aa2168503e8a"} Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.146409 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57ec8f7742f7bac69f31a4a1ab85d6d2e2887f56af4c7d88fc98aa2168503e8a" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.146286 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q9w8s" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.219254 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-k44x6"] Dec 03 19:08:04 crc kubenswrapper[4758]: E1203 19:08:04.220044 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" containerName="run-os-openstack-openstack-cell1" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.220133 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" containerName="run-os-openstack-openstack-cell1" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.220445 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" containerName="run-os-openstack-openstack-cell1" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.221339 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.229395 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-k44x6"] Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.255301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.255665 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45nc4\" (UniqueName: \"kubernetes.io/projected/9934ab2d-58c4-44e7-8021-b74ec84341f9-kube-api-access-45nc4\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.255866 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ceph\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.256111 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-inventory\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.358186 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key\") pod \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\" (UID: \"131aa418-2634-4ccf-8ae2-5ac72ecfc8eb\") " Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.358562 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ceph\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.358631 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-inventory\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.358735 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.358874 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45nc4\" (UniqueName: \"kubernetes.io/projected/9934ab2d-58c4-44e7-8021-b74ec84341f9-kube-api-access-45nc4\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.363053 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb" (UID: "131aa418-2634-4ccf-8ae2-5ac72ecfc8eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.363105 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.363244 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-inventory\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.366700 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ceph\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.379094 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45nc4\" (UniqueName: \"kubernetes.io/projected/9934ab2d-58c4-44e7-8021-b74ec84341f9-kube-api-access-45nc4\") pod \"reboot-os-openstack-openstack-cell1-k44x6\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.461969 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/131aa418-2634-4ccf-8ae2-5ac72ecfc8eb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:04 crc kubenswrapper[4758]: I1203 19:08:04.545476 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:05 crc kubenswrapper[4758]: I1203 19:08:05.129118 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-k44x6"] Dec 03 19:08:05 crc kubenswrapper[4758]: I1203 19:08:05.158431 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" event={"ID":"9934ab2d-58c4-44e7-8021-b74ec84341f9","Type":"ContainerStarted","Data":"3a07270514d481f728c899f81a4c3998f0260fb4ae63b4676475e2fdc9c4dae9"} Dec 03 19:08:06 crc kubenswrapper[4758]: I1203 19:08:06.169396 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" event={"ID":"9934ab2d-58c4-44e7-8021-b74ec84341f9","Type":"ContainerStarted","Data":"423055cf295c2164ab210209e4ee0723ecb3bbd46704e38df87d161b3d01684b"} Dec 03 19:08:06 crc kubenswrapper[4758]: I1203 19:08:06.197104 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" podStartSLOduration=2.014817626 podStartE2EDuration="2.197084212s" podCreationTimestamp="2025-12-03 19:08:04 +0000 UTC" firstStartedPulling="2025-12-03 19:08:05.121754164 +0000 UTC m=+7940.323131025" lastFinishedPulling="2025-12-03 19:08:05.30402075 +0000 UTC m=+7940.505397611" observedRunningTime="2025-12-03 19:08:06.194205243 +0000 UTC m=+7941.395582114" watchObservedRunningTime="2025-12-03 19:08:06.197084212 +0000 UTC m=+7941.398461073" Dec 03 19:08:09 crc kubenswrapper[4758]: I1203 19:08:09.115040 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:08:09 crc kubenswrapper[4758]: E1203 19:08:09.115925 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:08:20 crc kubenswrapper[4758]: E1203 19:08:20.820352 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9934ab2d_58c4_44e7_8021_b74ec84341f9.slice/crio-423055cf295c2164ab210209e4ee0723ecb3bbd46704e38df87d161b3d01684b.scope\": RecentStats: unable to find data in memory cache]" Dec 03 19:08:21 crc kubenswrapper[4758]: I1203 19:08:21.314334 4758 generic.go:334] "Generic (PLEG): container finished" podID="9934ab2d-58c4-44e7-8021-b74ec84341f9" containerID="423055cf295c2164ab210209e4ee0723ecb3bbd46704e38df87d161b3d01684b" exitCode=0 Dec 03 19:08:21 crc kubenswrapper[4758]: I1203 19:08:21.314668 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" event={"ID":"9934ab2d-58c4-44e7-8021-b74ec84341f9","Type":"ContainerDied","Data":"423055cf295c2164ab210209e4ee0723ecb3bbd46704e38df87d161b3d01684b"} Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.764340 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.886768 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45nc4\" (UniqueName: \"kubernetes.io/projected/9934ab2d-58c4-44e7-8021-b74ec84341f9-kube-api-access-45nc4\") pod \"9934ab2d-58c4-44e7-8021-b74ec84341f9\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.886840 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-inventory\") pod \"9934ab2d-58c4-44e7-8021-b74ec84341f9\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.886863 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ssh-key\") pod \"9934ab2d-58c4-44e7-8021-b74ec84341f9\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.886946 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ceph\") pod \"9934ab2d-58c4-44e7-8021-b74ec84341f9\" (UID: \"9934ab2d-58c4-44e7-8021-b74ec84341f9\") " Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.892897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9934ab2d-58c4-44e7-8021-b74ec84341f9-kube-api-access-45nc4" (OuterVolumeSpecName: "kube-api-access-45nc4") pod "9934ab2d-58c4-44e7-8021-b74ec84341f9" (UID: "9934ab2d-58c4-44e7-8021-b74ec84341f9"). InnerVolumeSpecName "kube-api-access-45nc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.896196 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ceph" (OuterVolumeSpecName: "ceph") pod "9934ab2d-58c4-44e7-8021-b74ec84341f9" (UID: "9934ab2d-58c4-44e7-8021-b74ec84341f9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.917771 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9934ab2d-58c4-44e7-8021-b74ec84341f9" (UID: "9934ab2d-58c4-44e7-8021-b74ec84341f9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.923942 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-inventory" (OuterVolumeSpecName: "inventory") pod "9934ab2d-58c4-44e7-8021-b74ec84341f9" (UID: "9934ab2d-58c4-44e7-8021-b74ec84341f9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.988914 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45nc4\" (UniqueName: \"kubernetes.io/projected/9934ab2d-58c4-44e7-8021-b74ec84341f9-kube-api-access-45nc4\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.988947 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.988956 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:22 crc kubenswrapper[4758]: I1203 19:08:22.988965 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9934ab2d-58c4-44e7-8021-b74ec84341f9-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.336307 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" event={"ID":"9934ab2d-58c4-44e7-8021-b74ec84341f9","Type":"ContainerDied","Data":"3a07270514d481f728c899f81a4c3998f0260fb4ae63b4676475e2fdc9c4dae9"} Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.336354 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a07270514d481f728c899f81a4c3998f0260fb4ae63b4676475e2fdc9c4dae9" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.336359 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-k44x6" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.420344 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-wp698"] Dec 03 19:08:23 crc kubenswrapper[4758]: E1203 19:08:23.420893 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9934ab2d-58c4-44e7-8021-b74ec84341f9" containerName="reboot-os-openstack-openstack-cell1" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.420914 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9934ab2d-58c4-44e7-8021-b74ec84341f9" containerName="reboot-os-openstack-openstack-cell1" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.421214 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9934ab2d-58c4-44e7-8021-b74ec84341f9" containerName="reboot-os-openstack-openstack-cell1" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.422164 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.424297 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.424563 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.424832 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.424953 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.444993 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-wp698"] Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.498855 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499090 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499152 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ssh-key\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ceph\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499430 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv2p5\" (UniqueName: \"kubernetes.io/projected/29991f51-5ebf-4690-9196-7bb5a47202e9-kube-api-access-lv2p5\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-inventory\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499586 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499666 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499777 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.499960 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.500000 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602095 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602186 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602209 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602242 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602325 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602346 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ssh-key\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602389 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ceph\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602408 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv2p5\" (UniqueName: \"kubernetes.io/projected/29991f51-5ebf-4690-9196-7bb5a47202e9-kube-api-access-lv2p5\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602434 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602452 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-inventory\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602482 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.602503 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.606823 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ceph\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.606942 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.607852 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-inventory\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.606831 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.608152 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.608563 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.608750 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.608776 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.609655 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.610194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.616202 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ssh-key\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.621266 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv2p5\" (UniqueName: \"kubernetes.io/projected/29991f51-5ebf-4690-9196-7bb5a47202e9-kube-api-access-lv2p5\") pod \"install-certs-openstack-openstack-cell1-wp698\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:23 crc kubenswrapper[4758]: I1203 19:08:23.747450 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:24 crc kubenswrapper[4758]: I1203 19:08:24.114329 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:08:24 crc kubenswrapper[4758]: E1203 19:08:24.114918 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:08:24 crc kubenswrapper[4758]: I1203 19:08:24.275107 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-wp698"] Dec 03 19:08:24 crc kubenswrapper[4758]: I1203 19:08:24.346608 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wp698" event={"ID":"29991f51-5ebf-4690-9196-7bb5a47202e9","Type":"ContainerStarted","Data":"49cd10d1a7b1580f19154b648cb5b54336d672e53994232ed3c23e7201705669"} Dec 03 19:08:25 crc kubenswrapper[4758]: I1203 19:08:25.357667 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wp698" event={"ID":"29991f51-5ebf-4690-9196-7bb5a47202e9","Type":"ContainerStarted","Data":"27fd8534b12a61ee25608398956ac818a3394d9eeeb2943f5843284dbdc1953e"} Dec 03 19:08:25 crc kubenswrapper[4758]: I1203 19:08:25.388830 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-wp698" podStartSLOduration=2.210523776 podStartE2EDuration="2.388811914s" podCreationTimestamp="2025-12-03 19:08:23 +0000 UTC" firstStartedPulling="2025-12-03 19:08:24.281653801 +0000 UTC m=+7959.483030662" lastFinishedPulling="2025-12-03 19:08:24.459941939 +0000 UTC m=+7959.661318800" observedRunningTime="2025-12-03 19:08:25.378420962 +0000 UTC m=+7960.579797833" watchObservedRunningTime="2025-12-03 19:08:25.388811914 +0000 UTC m=+7960.590188765" Dec 03 19:08:38 crc kubenswrapper[4758]: I1203 19:08:38.115055 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:08:38 crc kubenswrapper[4758]: E1203 19:08:38.116058 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:08:44 crc kubenswrapper[4758]: I1203 19:08:44.559241 4758 generic.go:334] "Generic (PLEG): container finished" podID="29991f51-5ebf-4690-9196-7bb5a47202e9" containerID="27fd8534b12a61ee25608398956ac818a3394d9eeeb2943f5843284dbdc1953e" exitCode=0 Dec 03 19:08:44 crc kubenswrapper[4758]: I1203 19:08:44.559442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wp698" event={"ID":"29991f51-5ebf-4690-9196-7bb5a47202e9","Type":"ContainerDied","Data":"27fd8534b12a61ee25608398956ac818a3394d9eeeb2943f5843284dbdc1953e"} Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.052515 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.156965 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv2p5\" (UniqueName: \"kubernetes.io/projected/29991f51-5ebf-4690-9196-7bb5a47202e9-kube-api-access-lv2p5\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157096 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-nova-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157153 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ovn-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157253 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ceph\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157288 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-telemetry-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157392 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ssh-key\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157449 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-bootstrap-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157521 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-inventory\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-libvirt-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157605 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-metadata-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157632 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-sriov-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.157756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-dhcp-combined-ca-bundle\") pod \"29991f51-5ebf-4690-9196-7bb5a47202e9\" (UID: \"29991f51-5ebf-4690-9196-7bb5a47202e9\") " Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.164786 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.164829 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.164944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.165021 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.165096 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ceph" (OuterVolumeSpecName: "ceph") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.165258 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29991f51-5ebf-4690-9196-7bb5a47202e9-kube-api-access-lv2p5" (OuterVolumeSpecName: "kube-api-access-lv2p5") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "kube-api-access-lv2p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.168284 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.168266 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.168763 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.179851 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.195888 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-inventory" (OuterVolumeSpecName: "inventory") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.196860 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "29991f51-5ebf-4690-9196-7bb5a47202e9" (UID: "29991f51-5ebf-4690-9196-7bb5a47202e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.260868 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.260932 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv2p5\" (UniqueName: \"kubernetes.io/projected/29991f51-5ebf-4690-9196-7bb5a47202e9-kube-api-access-lv2p5\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.260947 4758 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.260961 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.260975 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.260989 4758 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.261000 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.261012 4758 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.261022 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.261034 4758 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.261044 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.261056 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29991f51-5ebf-4690-9196-7bb5a47202e9-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.582161 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wp698" event={"ID":"29991f51-5ebf-4690-9196-7bb5a47202e9","Type":"ContainerDied","Data":"49cd10d1a7b1580f19154b648cb5b54336d672e53994232ed3c23e7201705669"} Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.582477 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49cd10d1a7b1580f19154b648cb5b54336d672e53994232ed3c23e7201705669" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.582274 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wp698" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.688104 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-649l7"] Dec 03 19:08:46 crc kubenswrapper[4758]: E1203 19:08:46.688720 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29991f51-5ebf-4690-9196-7bb5a47202e9" containerName="install-certs-openstack-openstack-cell1" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.688741 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="29991f51-5ebf-4690-9196-7bb5a47202e9" containerName="install-certs-openstack-openstack-cell1" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.688947 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="29991f51-5ebf-4690-9196-7bb5a47202e9" containerName="install-certs-openstack-openstack-cell1" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.689803 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.692116 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.693928 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.693955 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.694070 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.699746 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-649l7"] Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.773474 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk4w2\" (UniqueName: \"kubernetes.io/projected/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-kube-api-access-fk4w2\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.773656 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-inventory\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.774038 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ceph\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.774118 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.875775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ceph\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.875821 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.875945 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk4w2\" (UniqueName: \"kubernetes.io/projected/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-kube-api-access-fk4w2\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.875998 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-inventory\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.880947 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ceph\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.888292 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.888300 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-inventory\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:46 crc kubenswrapper[4758]: I1203 19:08:46.893608 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk4w2\" (UniqueName: \"kubernetes.io/projected/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-kube-api-access-fk4w2\") pod \"ceph-client-openstack-openstack-cell1-649l7\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:47 crc kubenswrapper[4758]: I1203 19:08:47.014950 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:47 crc kubenswrapper[4758]: I1203 19:08:47.595926 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-649l7"] Dec 03 19:08:48 crc kubenswrapper[4758]: I1203 19:08:48.606456 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" event={"ID":"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92","Type":"ContainerStarted","Data":"fade8710671c1aa7fbb071d640bf21ad471300ecca1a382b218d77a9bc73bb70"} Dec 03 19:08:48 crc kubenswrapper[4758]: I1203 19:08:48.607019 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" event={"ID":"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92","Type":"ContainerStarted","Data":"27b88b5de45165c172f69f69550ef26547ee7881b10573197dbfe34f8f659c0e"} Dec 03 19:08:48 crc kubenswrapper[4758]: I1203 19:08:48.630399 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" podStartSLOduration=2.4493758420000002 podStartE2EDuration="2.630384153s" podCreationTimestamp="2025-12-03 19:08:46 +0000 UTC" firstStartedPulling="2025-12-03 19:08:47.596987625 +0000 UTC m=+7982.798364486" lastFinishedPulling="2025-12-03 19:08:47.777995906 +0000 UTC m=+7982.979372797" observedRunningTime="2025-12-03 19:08:48.629272422 +0000 UTC m=+7983.830649283" watchObservedRunningTime="2025-12-03 19:08:48.630384153 +0000 UTC m=+7983.831761014" Dec 03 19:08:50 crc kubenswrapper[4758]: I1203 19:08:50.114795 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:08:50 crc kubenswrapper[4758]: E1203 19:08:50.115415 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:08:53 crc kubenswrapper[4758]: I1203 19:08:53.670046 4758 generic.go:334] "Generic (PLEG): container finished" podID="bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" containerID="fade8710671c1aa7fbb071d640bf21ad471300ecca1a382b218d77a9bc73bb70" exitCode=0 Dec 03 19:08:53 crc kubenswrapper[4758]: I1203 19:08:53.670786 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" event={"ID":"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92","Type":"ContainerDied","Data":"fade8710671c1aa7fbb071d640bf21ad471300ecca1a382b218d77a9bc73bb70"} Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.150293 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.262489 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-inventory\") pod \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.262590 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ssh-key\") pod \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.262618 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ceph\") pod \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.262958 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk4w2\" (UniqueName: \"kubernetes.io/projected/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-kube-api-access-fk4w2\") pod \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\" (UID: \"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92\") " Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.277253 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-kube-api-access-fk4w2" (OuterVolumeSpecName: "kube-api-access-fk4w2") pod "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" (UID: "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92"). InnerVolumeSpecName "kube-api-access-fk4w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.283849 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ceph" (OuterVolumeSpecName: "ceph") pod "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" (UID: "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.301981 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-inventory" (OuterVolumeSpecName: "inventory") pod "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" (UID: "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.317237 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" (UID: "bf9c857e-ddc6-41fb-bc6c-9fe40977bc92"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.366217 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.366254 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.366267 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk4w2\" (UniqueName: \"kubernetes.io/projected/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-kube-api-access-fk4w2\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.366282 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9c857e-ddc6-41fb-bc6c-9fe40977bc92-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.693571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" event={"ID":"bf9c857e-ddc6-41fb-bc6c-9fe40977bc92","Type":"ContainerDied","Data":"27b88b5de45165c172f69f69550ef26547ee7881b10573197dbfe34f8f659c0e"} Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.693619 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27b88b5de45165c172f69f69550ef26547ee7881b10573197dbfe34f8f659c0e" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.693716 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-649l7" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.784296 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-6mrsp"] Dec 03 19:08:55 crc kubenswrapper[4758]: E1203 19:08:55.785009 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" containerName="ceph-client-openstack-openstack-cell1" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.785082 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" containerName="ceph-client-openstack-openstack-cell1" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.785342 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9c857e-ddc6-41fb-bc6c-9fe40977bc92" containerName="ceph-client-openstack-openstack-cell1" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.786288 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.788400 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.789197 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.789433 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.789868 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.789910 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.795779 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-6mrsp"] Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.877185 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-inventory\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.877254 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.877295 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ceph\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.877328 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ssh-key\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.877388 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk44w\" (UniqueName: \"kubernetes.io/projected/92c8b34e-4a7c-49ad-8678-b9d85206ee55-kube-api-access-dk44w\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.877490 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.979674 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk44w\" (UniqueName: \"kubernetes.io/projected/92c8b34e-4a7c-49ad-8678-b9d85206ee55-kube-api-access-dk44w\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.979792 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.979888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-inventory\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.979915 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.979940 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ceph\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.979960 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ssh-key\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.981612 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.983937 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ssh-key\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.984350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-inventory\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.984805 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ceph\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:55 crc kubenswrapper[4758]: I1203 19:08:55.999594 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk44w\" (UniqueName: \"kubernetes.io/projected/92c8b34e-4a7c-49ad-8678-b9d85206ee55-kube-api-access-dk44w\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:56 crc kubenswrapper[4758]: I1203 19:08:56.000743 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-6mrsp\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:56 crc kubenswrapper[4758]: I1203 19:08:56.103871 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:08:56 crc kubenswrapper[4758]: I1203 19:08:56.708565 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-6mrsp"] Dec 03 19:08:57 crc kubenswrapper[4758]: I1203 19:08:57.716375 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" event={"ID":"92c8b34e-4a7c-49ad-8678-b9d85206ee55","Type":"ContainerStarted","Data":"38b16ab6c3b423e6937ac3a40241161416edc4971a8254669587c21d25cf32e0"} Dec 03 19:08:57 crc kubenswrapper[4758]: I1203 19:08:57.716718 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" event={"ID":"92c8b34e-4a7c-49ad-8678-b9d85206ee55","Type":"ContainerStarted","Data":"ac9a8cd56675ab57b5d9820e8bca396e44da19390be84bb8d67dca28588e1dad"} Dec 03 19:08:57 crc kubenswrapper[4758]: I1203 19:08:57.744375 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" podStartSLOduration=2.56593998 podStartE2EDuration="2.744348451s" podCreationTimestamp="2025-12-03 19:08:55 +0000 UTC" firstStartedPulling="2025-12-03 19:08:56.726835445 +0000 UTC m=+7991.928212346" lastFinishedPulling="2025-12-03 19:08:56.905243956 +0000 UTC m=+7992.106620817" observedRunningTime="2025-12-03 19:08:57.74173934 +0000 UTC m=+7992.943116211" watchObservedRunningTime="2025-12-03 19:08:57.744348451 +0000 UTC m=+7992.945725342" Dec 03 19:09:04 crc kubenswrapper[4758]: I1203 19:09:04.114293 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:09:04 crc kubenswrapper[4758]: E1203 19:09:04.115906 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:09:16 crc kubenswrapper[4758]: I1203 19:09:16.115203 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:09:16 crc kubenswrapper[4758]: E1203 19:09:16.116483 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:09:30 crc kubenswrapper[4758]: I1203 19:09:30.115082 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:09:30 crc kubenswrapper[4758]: E1203 19:09:30.116078 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:09:43 crc kubenswrapper[4758]: I1203 19:09:43.114107 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:09:43 crc kubenswrapper[4758]: E1203 19:09:43.115044 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:09:56 crc kubenswrapper[4758]: I1203 19:09:56.114380 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:09:56 crc kubenswrapper[4758]: E1203 19:09:56.115480 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:10:03 crc kubenswrapper[4758]: I1203 19:10:03.413937 4758 generic.go:334] "Generic (PLEG): container finished" podID="92c8b34e-4a7c-49ad-8678-b9d85206ee55" containerID="38b16ab6c3b423e6937ac3a40241161416edc4971a8254669587c21d25cf32e0" exitCode=0 Dec 03 19:10:03 crc kubenswrapper[4758]: I1203 19:10:03.414007 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" event={"ID":"92c8b34e-4a7c-49ad-8678-b9d85206ee55","Type":"ContainerDied","Data":"38b16ab6c3b423e6937ac3a40241161416edc4971a8254669587c21d25cf32e0"} Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.872674 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.993943 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-inventory\") pod \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.994029 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovn-combined-ca-bundle\") pod \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.994166 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk44w\" (UniqueName: \"kubernetes.io/projected/92c8b34e-4a7c-49ad-8678-b9d85206ee55-kube-api-access-dk44w\") pod \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.994201 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovncontroller-config-0\") pod \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.994279 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ssh-key\") pod \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " Dec 03 19:10:04 crc kubenswrapper[4758]: I1203 19:10:04.994322 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ceph\") pod \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\" (UID: \"92c8b34e-4a7c-49ad-8678-b9d85206ee55\") " Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.001657 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c8b34e-4a7c-49ad-8678-b9d85206ee55-kube-api-access-dk44w" (OuterVolumeSpecName: "kube-api-access-dk44w") pod "92c8b34e-4a7c-49ad-8678-b9d85206ee55" (UID: "92c8b34e-4a7c-49ad-8678-b9d85206ee55"). InnerVolumeSpecName "kube-api-access-dk44w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.001971 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "92c8b34e-4a7c-49ad-8678-b9d85206ee55" (UID: "92c8b34e-4a7c-49ad-8678-b9d85206ee55"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.009442 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ceph" (OuterVolumeSpecName: "ceph") pod "92c8b34e-4a7c-49ad-8678-b9d85206ee55" (UID: "92c8b34e-4a7c-49ad-8678-b9d85206ee55"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.026235 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "92c8b34e-4a7c-49ad-8678-b9d85206ee55" (UID: "92c8b34e-4a7c-49ad-8678-b9d85206ee55"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.031225 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-inventory" (OuterVolumeSpecName: "inventory") pod "92c8b34e-4a7c-49ad-8678-b9d85206ee55" (UID: "92c8b34e-4a7c-49ad-8678-b9d85206ee55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.049846 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92c8b34e-4a7c-49ad-8678-b9d85206ee55" (UID: "92c8b34e-4a7c-49ad-8678-b9d85206ee55"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.098516 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.098585 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.098606 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk44w\" (UniqueName: \"kubernetes.io/projected/92c8b34e-4a7c-49ad-8678-b9d85206ee55-kube-api-access-dk44w\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.098620 4758 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.098634 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.098664 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92c8b34e-4a7c-49ad-8678-b9d85206ee55-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.439228 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" event={"ID":"92c8b34e-4a7c-49ad-8678-b9d85206ee55","Type":"ContainerDied","Data":"ac9a8cd56675ab57b5d9820e8bca396e44da19390be84bb8d67dca28588e1dad"} Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.439267 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac9a8cd56675ab57b5d9820e8bca396e44da19390be84bb8d67dca28588e1dad" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.439309 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6mrsp" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.543917 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xrh25"] Dec 03 19:10:05 crc kubenswrapper[4758]: E1203 19:10:05.544496 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c8b34e-4a7c-49ad-8678-b9d85206ee55" containerName="ovn-openstack-openstack-cell1" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.544521 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c8b34e-4a7c-49ad-8678-b9d85206ee55" containerName="ovn-openstack-openstack-cell1" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.544760 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c8b34e-4a7c-49ad-8678-b9d85206ee55" containerName="ovn-openstack-openstack-cell1" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.545466 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.556616 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xrh25"] Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.558982 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.559407 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.559407 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.559407 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.559499 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.559614 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609305 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609353 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgr8f\" (UniqueName: \"kubernetes.io/projected/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-kube-api-access-pgr8f\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609421 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609495 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609588 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.609946 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.712340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.713009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.713115 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.713171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.713302 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.713343 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgr8f\" (UniqueName: \"kubernetes.io/projected/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-kube-api-access-pgr8f\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.713426 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.717395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.717425 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.717841 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.717849 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.718259 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.725149 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.742949 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgr8f\" (UniqueName: \"kubernetes.io/projected/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-kube-api-access-pgr8f\") pod \"neutron-metadata-openstack-openstack-cell1-xrh25\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:05 crc kubenswrapper[4758]: I1203 19:10:05.875191 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:10:06 crc kubenswrapper[4758]: I1203 19:10:06.648715 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xrh25"] Dec 03 19:10:06 crc kubenswrapper[4758]: I1203 19:10:06.653318 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:10:07 crc kubenswrapper[4758]: I1203 19:10:07.457946 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" event={"ID":"1eb85ae0-647a-4e35-a8bb-404f07ee1eee","Type":"ContainerStarted","Data":"f06e41a547460fae1f211d5479e71d576ae1cee2f9bf2eeb5dab77c5e46f7f1d"} Dec 03 19:10:07 crc kubenswrapper[4758]: I1203 19:10:07.458245 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" event={"ID":"1eb85ae0-647a-4e35-a8bb-404f07ee1eee","Type":"ContainerStarted","Data":"ae08f788daea1a71a99d2e7bdd2f7ecb99ecb890eee626490035c58cafc82606"} Dec 03 19:10:07 crc kubenswrapper[4758]: I1203 19:10:07.490562 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" podStartSLOduration=2.306391684 podStartE2EDuration="2.49053852s" podCreationTimestamp="2025-12-03 19:10:05 +0000 UTC" firstStartedPulling="2025-12-03 19:10:06.653090696 +0000 UTC m=+8061.854467547" lastFinishedPulling="2025-12-03 19:10:06.837237522 +0000 UTC m=+8062.038614383" observedRunningTime="2025-12-03 19:10:07.480083537 +0000 UTC m=+8062.681460398" watchObservedRunningTime="2025-12-03 19:10:07.49053852 +0000 UTC m=+8062.691915381" Dec 03 19:10:10 crc kubenswrapper[4758]: I1203 19:10:10.114880 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:10:10 crc kubenswrapper[4758]: E1203 19:10:10.115630 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:10:24 crc kubenswrapper[4758]: I1203 19:10:24.117632 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:10:24 crc kubenswrapper[4758]: E1203 19:10:24.118386 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:10:37 crc kubenswrapper[4758]: I1203 19:10:37.114186 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:10:37 crc kubenswrapper[4758]: E1203 19:10:37.115240 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:10:50 crc kubenswrapper[4758]: I1203 19:10:50.114556 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:10:50 crc kubenswrapper[4758]: E1203 19:10:50.115446 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:11:02 crc kubenswrapper[4758]: I1203 19:11:02.077462 4758 generic.go:334] "Generic (PLEG): container finished" podID="1eb85ae0-647a-4e35-a8bb-404f07ee1eee" containerID="f06e41a547460fae1f211d5479e71d576ae1cee2f9bf2eeb5dab77c5e46f7f1d" exitCode=0 Dec 03 19:11:02 crc kubenswrapper[4758]: I1203 19:11:02.077549 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" event={"ID":"1eb85ae0-647a-4e35-a8bb-404f07ee1eee","Type":"ContainerDied","Data":"f06e41a547460fae1f211d5479e71d576ae1cee2f9bf2eeb5dab77c5e46f7f1d"} Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.590319 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692232 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-metadata-combined-ca-bundle\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgr8f\" (UniqueName: \"kubernetes.io/projected/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-kube-api-access-pgr8f\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ceph\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692601 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692633 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-nova-metadata-neutron-config-0\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692777 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-inventory\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.692820 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.706722 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.707099 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-kube-api-access-pgr8f" (OuterVolumeSpecName: "kube-api-access-pgr8f") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "kube-api-access-pgr8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.708766 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ceph" (OuterVolumeSpecName: "ceph") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:11:03 crc kubenswrapper[4758]: E1203 19:11:03.727975 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key podName:1eb85ae0-647a-4e35-a8bb-404f07ee1eee nodeName:}" failed. No retries permitted until 2025-12-03 19:11:04.227873373 +0000 UTC m=+8119.429250234 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee") : error deleting /var/lib/kubelet/pods/1eb85ae0-647a-4e35-a8bb-404f07ee1eee/volume-subpaths: remove /var/lib/kubelet/pods/1eb85ae0-647a-4e35-a8bb-404f07ee1eee/volume-subpaths: no such file or directory Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.729163 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.729617 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-inventory" (OuterVolumeSpecName: "inventory") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.730500 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.795810 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.795883 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.795897 4758 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.795907 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.795915 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:03 crc kubenswrapper[4758]: I1203 19:11:03.795946 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgr8f\" (UniqueName: \"kubernetes.io/projected/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-kube-api-access-pgr8f\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.095987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" event={"ID":"1eb85ae0-647a-4e35-a8bb-404f07ee1eee","Type":"ContainerDied","Data":"ae08f788daea1a71a99d2e7bdd2f7ecb99ecb890eee626490035c58cafc82606"} Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.096265 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae08f788daea1a71a99d2e7bdd2f7ecb99ecb890eee626490035c58cafc82606" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.096086 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xrh25" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.192738 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-wkqhj"] Dec 03 19:11:04 crc kubenswrapper[4758]: E1203 19:11:04.193260 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb85ae0-647a-4e35-a8bb-404f07ee1eee" containerName="neutron-metadata-openstack-openstack-cell1" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.193289 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb85ae0-647a-4e35-a8bb-404f07ee1eee" containerName="neutron-metadata-openstack-openstack-cell1" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.193517 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb85ae0-647a-4e35-a8bb-404f07ee1eee" containerName="neutron-metadata-openstack-openstack-cell1" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.194584 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.196599 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.216001 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-wkqhj"] Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.305047 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key\") pod \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\" (UID: \"1eb85ae0-647a-4e35-a8bb-404f07ee1eee\") " Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.306250 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.306452 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.306740 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ssh-key\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.306878 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ceph\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.306987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8t7v\" (UniqueName: \"kubernetes.io/projected/0594107d-92c8-4f63-9d64-5c362703c627-kube-api-access-q8t7v\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.307262 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-inventory\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.319212 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1eb85ae0-647a-4e35-a8bb-404f07ee1eee" (UID: "1eb85ae0-647a-4e35-a8bb-404f07ee1eee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409190 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409246 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ssh-key\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ceph\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409355 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8t7v\" (UniqueName: \"kubernetes.io/projected/0594107d-92c8-4f63-9d64-5c362703c627-kube-api-access-q8t7v\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409447 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-inventory\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.409511 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb85ae0-647a-4e35-a8bb-404f07ee1eee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.413211 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ssh-key\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.414096 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.414302 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-inventory\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.421370 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ceph\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.425318 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.430406 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8t7v\" (UniqueName: \"kubernetes.io/projected/0594107d-92c8-4f63-9d64-5c362703c627-kube-api-access-q8t7v\") pod \"libvirt-openstack-openstack-cell1-wkqhj\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:04 crc kubenswrapper[4758]: I1203 19:11:04.515912 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:11:05 crc kubenswrapper[4758]: W1203 19:11:05.045291 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0594107d_92c8_4f63_9d64_5c362703c627.slice/crio-85232d39884e3013295752f1555aa31ee9696180401484d18cbce827b7587073 WatchSource:0}: Error finding container 85232d39884e3013295752f1555aa31ee9696180401484d18cbce827b7587073: Status 404 returned error can't find the container with id 85232d39884e3013295752f1555aa31ee9696180401484d18cbce827b7587073 Dec 03 19:11:05 crc kubenswrapper[4758]: I1203 19:11:05.052938 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-wkqhj"] Dec 03 19:11:05 crc kubenswrapper[4758]: I1203 19:11:05.107513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" event={"ID":"0594107d-92c8-4f63-9d64-5c362703c627","Type":"ContainerStarted","Data":"85232d39884e3013295752f1555aa31ee9696180401484d18cbce827b7587073"} Dec 03 19:11:05 crc kubenswrapper[4758]: I1203 19:11:05.122309 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:11:05 crc kubenswrapper[4758]: E1203 19:11:05.122628 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:11:06 crc kubenswrapper[4758]: I1203 19:11:06.120768 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" event={"ID":"0594107d-92c8-4f63-9d64-5c362703c627","Type":"ContainerStarted","Data":"2da5a80861307b8ceba5d5d38d88f539ac7c3d0455664bc129e9295036dcc5fb"} Dec 03 19:11:06 crc kubenswrapper[4758]: I1203 19:11:06.143720 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" podStartSLOduration=1.965610825 podStartE2EDuration="2.143698048s" podCreationTimestamp="2025-12-03 19:11:04 +0000 UTC" firstStartedPulling="2025-12-03 19:11:05.047794048 +0000 UTC m=+8120.249170909" lastFinishedPulling="2025-12-03 19:11:05.225881271 +0000 UTC m=+8120.427258132" observedRunningTime="2025-12-03 19:11:06.135312721 +0000 UTC m=+8121.336689602" watchObservedRunningTime="2025-12-03 19:11:06.143698048 +0000 UTC m=+8121.345074929" Dec 03 19:11:20 crc kubenswrapper[4758]: I1203 19:11:20.115849 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:11:20 crc kubenswrapper[4758]: E1203 19:11:20.117303 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:11:32 crc kubenswrapper[4758]: I1203 19:11:32.114177 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:11:32 crc kubenswrapper[4758]: E1203 19:11:32.115014 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:11:43 crc kubenswrapper[4758]: I1203 19:11:43.115119 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:11:43 crc kubenswrapper[4758]: E1203 19:11:43.115995 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:11:56 crc kubenswrapper[4758]: I1203 19:11:56.114408 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:11:56 crc kubenswrapper[4758]: E1203 19:11:56.115298 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:12:10 crc kubenswrapper[4758]: I1203 19:12:10.114880 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:12:10 crc kubenswrapper[4758]: E1203 19:12:10.115569 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:12:23 crc kubenswrapper[4758]: I1203 19:12:23.114604 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:12:23 crc kubenswrapper[4758]: E1203 19:12:23.115658 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:12:36 crc kubenswrapper[4758]: I1203 19:12:36.114465 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:12:36 crc kubenswrapper[4758]: E1203 19:12:36.115424 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:12:49 crc kubenswrapper[4758]: I1203 19:12:49.114399 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:12:50 crc kubenswrapper[4758]: I1203 19:12:50.318953 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"435e139b7660730a3057119bb44eca7c2106ea848a8a6724b9ce30a20679541a"} Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.711920 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-87tgl"] Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.716623 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.741234 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-87tgl"] Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.785670 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtwvb\" (UniqueName: \"kubernetes.io/projected/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-kube-api-access-xtwvb\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.785818 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-catalog-content\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.785851 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-utilities\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.888191 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-catalog-content\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.888254 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-utilities\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.888406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtwvb\" (UniqueName: \"kubernetes.io/projected/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-kube-api-access-xtwvb\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.888893 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-catalog-content\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.888962 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-utilities\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:26 crc kubenswrapper[4758]: I1203 19:14:26.913495 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtwvb\" (UniqueName: \"kubernetes.io/projected/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-kube-api-access-xtwvb\") pod \"redhat-operators-87tgl\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:27 crc kubenswrapper[4758]: I1203 19:14:27.041884 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:27 crc kubenswrapper[4758]: I1203 19:14:27.551020 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-87tgl"] Dec 03 19:14:28 crc kubenswrapper[4758]: I1203 19:14:28.318648 4758 generic.go:334] "Generic (PLEG): container finished" podID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerID="c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d" exitCode=0 Dec 03 19:14:28 crc kubenswrapper[4758]: I1203 19:14:28.318717 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerDied","Data":"c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d"} Dec 03 19:14:28 crc kubenswrapper[4758]: I1203 19:14:28.318979 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerStarted","Data":"279c57937d2997a921530bc9af6709fbc3f75dbca2ff57b1a6d05b81c754a756"} Dec 03 19:14:29 crc kubenswrapper[4758]: I1203 19:14:29.334615 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerStarted","Data":"33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d"} Dec 03 19:14:31 crc kubenswrapper[4758]: I1203 19:14:31.357812 4758 generic.go:334] "Generic (PLEG): container finished" podID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerID="33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d" exitCode=0 Dec 03 19:14:31 crc kubenswrapper[4758]: I1203 19:14:31.357890 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerDied","Data":"33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d"} Dec 03 19:14:32 crc kubenswrapper[4758]: I1203 19:14:32.375631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerStarted","Data":"c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6"} Dec 03 19:14:32 crc kubenswrapper[4758]: I1203 19:14:32.403558 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-87tgl" podStartSLOduration=2.963322292 podStartE2EDuration="6.403537794s" podCreationTimestamp="2025-12-03 19:14:26 +0000 UTC" firstStartedPulling="2025-12-03 19:14:28.321152425 +0000 UTC m=+8323.522529286" lastFinishedPulling="2025-12-03 19:14:31.761367927 +0000 UTC m=+8326.962744788" observedRunningTime="2025-12-03 19:14:32.403311748 +0000 UTC m=+8327.604688609" watchObservedRunningTime="2025-12-03 19:14:32.403537794 +0000 UTC m=+8327.604914655" Dec 03 19:14:37 crc kubenswrapper[4758]: I1203 19:14:37.042597 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:37 crc kubenswrapper[4758]: I1203 19:14:37.043289 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.091638 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-87tgl" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="registry-server" probeResult="failure" output=< Dec 03 19:14:38 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 19:14:38 crc kubenswrapper[4758]: > Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.684569 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-drz2t"] Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.686928 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.708490 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-drz2t"] Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.768880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-utilities\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.769232 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-catalog-content\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.769531 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkp6s\" (UniqueName: \"kubernetes.io/projected/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-kube-api-access-mkp6s\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.872005 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkp6s\" (UniqueName: \"kubernetes.io/projected/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-kube-api-access-mkp6s\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.872115 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-utilities\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.872236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-catalog-content\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.872713 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-utilities\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.872736 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-catalog-content\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:38 crc kubenswrapper[4758]: I1203 19:14:38.893725 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkp6s\" (UniqueName: \"kubernetes.io/projected/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-kube-api-access-mkp6s\") pod \"community-operators-drz2t\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:39 crc kubenswrapper[4758]: I1203 19:14:39.004228 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:39 crc kubenswrapper[4758]: W1203 19:14:39.605855 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5df89ac_08f7_4ad8_b0c3_e50b3beb0f02.slice/crio-0292b79a9c27bec32092ccf6223f7ff3093b54930f782a78b24b775fdaabed7e WatchSource:0}: Error finding container 0292b79a9c27bec32092ccf6223f7ff3093b54930f782a78b24b775fdaabed7e: Status 404 returned error can't find the container with id 0292b79a9c27bec32092ccf6223f7ff3093b54930f782a78b24b775fdaabed7e Dec 03 19:14:39 crc kubenswrapper[4758]: I1203 19:14:39.606172 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-drz2t"] Dec 03 19:14:40 crc kubenswrapper[4758]: I1203 19:14:40.451742 4758 generic.go:334] "Generic (PLEG): container finished" podID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerID="ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0" exitCode=0 Dec 03 19:14:40 crc kubenswrapper[4758]: I1203 19:14:40.452346 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerDied","Data":"ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0"} Dec 03 19:14:40 crc kubenswrapper[4758]: I1203 19:14:40.452376 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerStarted","Data":"0292b79a9c27bec32092ccf6223f7ff3093b54930f782a78b24b775fdaabed7e"} Dec 03 19:14:41 crc kubenswrapper[4758]: I1203 19:14:41.463229 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerStarted","Data":"5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6"} Dec 03 19:14:42 crc kubenswrapper[4758]: I1203 19:14:42.477413 4758 generic.go:334] "Generic (PLEG): container finished" podID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerID="5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6" exitCode=0 Dec 03 19:14:42 crc kubenswrapper[4758]: I1203 19:14:42.477469 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerDied","Data":"5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6"} Dec 03 19:14:43 crc kubenswrapper[4758]: I1203 19:14:43.492118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerStarted","Data":"ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2"} Dec 03 19:14:43 crc kubenswrapper[4758]: I1203 19:14:43.520237 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-drz2t" podStartSLOduration=3.114937575 podStartE2EDuration="5.520220505s" podCreationTimestamp="2025-12-03 19:14:38 +0000 UTC" firstStartedPulling="2025-12-03 19:14:40.454627899 +0000 UTC m=+8335.656004750" lastFinishedPulling="2025-12-03 19:14:42.859910819 +0000 UTC m=+8338.061287680" observedRunningTime="2025-12-03 19:14:43.516247889 +0000 UTC m=+8338.717624760" watchObservedRunningTime="2025-12-03 19:14:43.520220505 +0000 UTC m=+8338.721597356" Dec 03 19:14:47 crc kubenswrapper[4758]: I1203 19:14:47.094758 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:47 crc kubenswrapper[4758]: I1203 19:14:47.149966 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:47 crc kubenswrapper[4758]: I1203 19:14:47.331231 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-87tgl"] Dec 03 19:14:48 crc kubenswrapper[4758]: I1203 19:14:48.550443 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-87tgl" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="registry-server" containerID="cri-o://c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6" gracePeriod=2 Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.005392 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.005745 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.055254 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.060897 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.127782 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-utilities\") pod \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.127900 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtwvb\" (UniqueName: \"kubernetes.io/projected/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-kube-api-access-xtwvb\") pod \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.127981 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-catalog-content\") pod \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\" (UID: \"0b9adc27-43a2-4e4d-81e3-df427df9e0e4\") " Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.141542 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-utilities" (OuterVolumeSpecName: "utilities") pod "0b9adc27-43a2-4e4d-81e3-df427df9e0e4" (UID: "0b9adc27-43a2-4e4d-81e3-df427df9e0e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.196071 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-kube-api-access-xtwvb" (OuterVolumeSpecName: "kube-api-access-xtwvb") pod "0b9adc27-43a2-4e4d-81e3-df427df9e0e4" (UID: "0b9adc27-43a2-4e4d-81e3-df427df9e0e4"). InnerVolumeSpecName "kube-api-access-xtwvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.245153 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.258860 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtwvb\" (UniqueName: \"kubernetes.io/projected/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-kube-api-access-xtwvb\") on node \"crc\" DevicePath \"\"" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.363474 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b9adc27-43a2-4e4d-81e3-df427df9e0e4" (UID: "0b9adc27-43a2-4e4d-81e3-df427df9e0e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.463064 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b9adc27-43a2-4e4d-81e3-df427df9e0e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.561056 4758 generic.go:334] "Generic (PLEG): container finished" podID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerID="c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6" exitCode=0 Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.561138 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87tgl" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.561162 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerDied","Data":"c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6"} Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.561580 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87tgl" event={"ID":"0b9adc27-43a2-4e4d-81e3-df427df9e0e4","Type":"ContainerDied","Data":"279c57937d2997a921530bc9af6709fbc3f75dbca2ff57b1a6d05b81c754a756"} Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.561628 4758 scope.go:117] "RemoveContainer" containerID="c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.586810 4758 scope.go:117] "RemoveContainer" containerID="33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.595236 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-87tgl"] Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.607644 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-87tgl"] Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.615721 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.626592 4758 scope.go:117] "RemoveContainer" containerID="c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.662398 4758 scope.go:117] "RemoveContainer" containerID="c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6" Dec 03 19:14:49 crc kubenswrapper[4758]: E1203 19:14:49.663137 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6\": container with ID starting with c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6 not found: ID does not exist" containerID="c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.663211 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6"} err="failed to get container status \"c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6\": rpc error: code = NotFound desc = could not find container \"c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6\": container with ID starting with c014bb856d0a1d87ea09618c8f4c75a81e6bf74c6c444d9ed518521ce6590ab6 not found: ID does not exist" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.663285 4758 scope.go:117] "RemoveContainer" containerID="33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d" Dec 03 19:14:49 crc kubenswrapper[4758]: E1203 19:14:49.663821 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d\": container with ID starting with 33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d not found: ID does not exist" containerID="33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.663885 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d"} err="failed to get container status \"33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d\": rpc error: code = NotFound desc = could not find container \"33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d\": container with ID starting with 33ffa9e18a1a5abfd180ed1cece274b2734cd832d5dd409b231c6550c4a19a5d not found: ID does not exist" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.663917 4758 scope.go:117] "RemoveContainer" containerID="c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d" Dec 03 19:14:49 crc kubenswrapper[4758]: E1203 19:14:49.664455 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d\": container with ID starting with c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d not found: ID does not exist" containerID="c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d" Dec 03 19:14:49 crc kubenswrapper[4758]: I1203 19:14:49.664488 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d"} err="failed to get container status \"c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d\": rpc error: code = NotFound desc = could not find container \"c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d\": container with ID starting with c9322af8523fba441c73369cd42adefadefb66e865b12a16d5217a63b212d33d not found: ID does not exist" Dec 03 19:14:50 crc kubenswrapper[4758]: I1203 19:14:50.334555 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-drz2t"] Dec 03 19:14:51 crc kubenswrapper[4758]: I1203 19:14:51.130703 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" path="/var/lib/kubelet/pods/0b9adc27-43a2-4e4d-81e3-df427df9e0e4/volumes" Dec 03 19:14:51 crc kubenswrapper[4758]: I1203 19:14:51.580815 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-drz2t" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="registry-server" containerID="cri-o://ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2" gracePeriod=2 Dec 03 19:14:51 crc kubenswrapper[4758]: E1203 19:14:51.784524 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5df89ac_08f7_4ad8_b0c3_e50b3beb0f02.slice/crio-conmon-ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5df89ac_08f7_4ad8_b0c3_e50b3beb0f02.slice/crio-ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2.scope\": RecentStats: unable to find data in memory cache]" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.060703 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.115303 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-utilities\") pod \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.115989 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-catalog-content\") pod \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.116122 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkp6s\" (UniqueName: \"kubernetes.io/projected/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-kube-api-access-mkp6s\") pod \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\" (UID: \"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02\") " Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.120605 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-utilities" (OuterVolumeSpecName: "utilities") pod "d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" (UID: "d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.125283 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-kube-api-access-mkp6s" (OuterVolumeSpecName: "kube-api-access-mkp6s") pod "d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" (UID: "d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02"). InnerVolumeSpecName "kube-api-access-mkp6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.172943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" (UID: "d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.219554 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkp6s\" (UniqueName: \"kubernetes.io/projected/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-kube-api-access-mkp6s\") on node \"crc\" DevicePath \"\"" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.219602 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.219613 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.592136 4758 generic.go:334] "Generic (PLEG): container finished" podID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerID="ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2" exitCode=0 Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.592192 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerDied","Data":"ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2"} Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.592220 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drz2t" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.592238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drz2t" event={"ID":"d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02","Type":"ContainerDied","Data":"0292b79a9c27bec32092ccf6223f7ff3093b54930f782a78b24b775fdaabed7e"} Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.592257 4758 scope.go:117] "RemoveContainer" containerID="ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.624516 4758 scope.go:117] "RemoveContainer" containerID="5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.638597 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-drz2t"] Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.649646 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-drz2t"] Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.662965 4758 scope.go:117] "RemoveContainer" containerID="ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.708816 4758 scope.go:117] "RemoveContainer" containerID="ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.709357 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2\": container with ID starting with ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2 not found: ID does not exist" containerID="ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.709400 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2"} err="failed to get container status \"ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2\": rpc error: code = NotFound desc = could not find container \"ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2\": container with ID starting with ee8345ad277e8501100e049ae14127515d7e9d66dc9a42dd0fccd5e3a50faae2 not found: ID does not exist" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.709431 4758 scope.go:117] "RemoveContainer" containerID="5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.709937 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6\": container with ID starting with 5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6 not found: ID does not exist" containerID="5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.710008 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6"} err="failed to get container status \"5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6\": rpc error: code = NotFound desc = could not find container \"5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6\": container with ID starting with 5309e503beec9e2d10efc176503c9ee9acbce53525e5bde1767d1b541331c5d6 not found: ID does not exist" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.710061 4758 scope.go:117] "RemoveContainer" containerID="ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.710500 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0\": container with ID starting with ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0 not found: ID does not exist" containerID="ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.710529 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0"} err="failed to get container status \"ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0\": rpc error: code = NotFound desc = could not find container \"ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0\": container with ID starting with ee68aa2ffa00e869c98ee7686b11a096e041664b38c97747836d2979319c97f0 not found: ID does not exist" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.748373 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4lw69"] Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.749044 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="extract-content" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749067 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="extract-content" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.749093 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="registry-server" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749104 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="registry-server" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.749148 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="extract-utilities" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749161 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="extract-utilities" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.749197 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="registry-server" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749208 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="registry-server" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.749233 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="extract-utilities" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749243 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="extract-utilities" Dec 03 19:14:52 crc kubenswrapper[4758]: E1203 19:14:52.749276 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="extract-content" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749288 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="extract-content" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749629 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b9adc27-43a2-4e4d-81e3-df427df9e0e4" containerName="registry-server" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.749656 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" containerName="registry-server" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.752288 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.760939 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lw69"] Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.834727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-catalog-content\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.835002 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc4x6\" (UniqueName: \"kubernetes.io/projected/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-kube-api-access-xc4x6\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.835197 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-utilities\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.937903 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-catalog-content\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.937967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc4x6\" (UniqueName: \"kubernetes.io/projected/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-kube-api-access-xc4x6\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.938004 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-utilities\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.938510 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-catalog-content\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.938569 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-utilities\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:52 crc kubenswrapper[4758]: I1203 19:14:52.958366 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc4x6\" (UniqueName: \"kubernetes.io/projected/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-kube-api-access-xc4x6\") pod \"redhat-marketplace-4lw69\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:53 crc kubenswrapper[4758]: I1203 19:14:53.088276 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:14:53 crc kubenswrapper[4758]: I1203 19:14:53.127451 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02" path="/var/lib/kubelet/pods/d5df89ac-08f7-4ad8-b0c3-e50b3beb0f02/volumes" Dec 03 19:14:53 crc kubenswrapper[4758]: I1203 19:14:53.579979 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lw69"] Dec 03 19:14:53 crc kubenswrapper[4758]: I1203 19:14:53.605605 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lw69" event={"ID":"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4","Type":"ContainerStarted","Data":"74aa018f10bc939516e4b25e17b2d87610eed2c8a68125d5a54259f4d7d8194d"} Dec 03 19:14:54 crc kubenswrapper[4758]: I1203 19:14:54.620029 4758 generic.go:334] "Generic (PLEG): container finished" podID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerID="069b497884c33d9d7fc2ab60d56c474577bde955220fcbc2a45b81a43563f0fa" exitCode=0 Dec 03 19:14:54 crc kubenswrapper[4758]: I1203 19:14:54.620078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lw69" event={"ID":"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4","Type":"ContainerDied","Data":"069b497884c33d9d7fc2ab60d56c474577bde955220fcbc2a45b81a43563f0fa"} Dec 03 19:14:55 crc kubenswrapper[4758]: I1203 19:14:55.635986 4758 generic.go:334] "Generic (PLEG): container finished" podID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerID="a4ec28693c7b0a2ec9314b6e8c49c08b3316377de8df158427860055edb33b5f" exitCode=0 Dec 03 19:14:55 crc kubenswrapper[4758]: I1203 19:14:55.636077 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lw69" event={"ID":"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4","Type":"ContainerDied","Data":"a4ec28693c7b0a2ec9314b6e8c49c08b3316377de8df158427860055edb33b5f"} Dec 03 19:14:56 crc kubenswrapper[4758]: I1203 19:14:56.648988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lw69" event={"ID":"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4","Type":"ContainerStarted","Data":"48403a04660aceeebbe46dfff8d83bce23964a19d045a2bff0fa51c678ea3d43"} Dec 03 19:14:56 crc kubenswrapper[4758]: I1203 19:14:56.672233 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4lw69" podStartSLOduration=3.193128543 podStartE2EDuration="4.672215419s" podCreationTimestamp="2025-12-03 19:14:52 +0000 UTC" firstStartedPulling="2025-12-03 19:14:54.622317854 +0000 UTC m=+8349.823694705" lastFinishedPulling="2025-12-03 19:14:56.10140472 +0000 UTC m=+8351.302781581" observedRunningTime="2025-12-03 19:14:56.669438664 +0000 UTC m=+8351.870815525" watchObservedRunningTime="2025-12-03 19:14:56.672215419 +0000 UTC m=+8351.873592280" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.151550 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6"] Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.154023 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.158407 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.160441 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.165579 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6"] Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.217804 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aee032d-e890-4885-84ad-ce4dc9fbd20a-secret-volume\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.217905 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aee032d-e890-4885-84ad-ce4dc9fbd20a-config-volume\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.217964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfbrm\" (UniqueName: \"kubernetes.io/projected/2aee032d-e890-4885-84ad-ce4dc9fbd20a-kube-api-access-vfbrm\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.321368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aee032d-e890-4885-84ad-ce4dc9fbd20a-secret-volume\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.321897 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aee032d-e890-4885-84ad-ce4dc9fbd20a-config-volume\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.322041 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfbrm\" (UniqueName: \"kubernetes.io/projected/2aee032d-e890-4885-84ad-ce4dc9fbd20a-kube-api-access-vfbrm\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.322898 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aee032d-e890-4885-84ad-ce4dc9fbd20a-config-volume\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.333437 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aee032d-e890-4885-84ad-ce4dc9fbd20a-secret-volume\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.344515 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfbrm\" (UniqueName: \"kubernetes.io/projected/2aee032d-e890-4885-84ad-ce4dc9fbd20a-kube-api-access-vfbrm\") pod \"collect-profiles-29413155-4nlc6\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.480207 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:00 crc kubenswrapper[4758]: I1203 19:15:00.927404 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6"] Dec 03 19:15:00 crc kubenswrapper[4758]: W1203 19:15:00.936951 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2aee032d_e890_4885_84ad_ce4dc9fbd20a.slice/crio-0b72ecb560a437a96967488dba89887a15f0ca006f861f70fc1c8a7f7ae7daf4 WatchSource:0}: Error finding container 0b72ecb560a437a96967488dba89887a15f0ca006f861f70fc1c8a7f7ae7daf4: Status 404 returned error can't find the container with id 0b72ecb560a437a96967488dba89887a15f0ca006f861f70fc1c8a7f7ae7daf4 Dec 03 19:15:01 crc kubenswrapper[4758]: I1203 19:15:01.704833 4758 generic.go:334] "Generic (PLEG): container finished" podID="2aee032d-e890-4885-84ad-ce4dc9fbd20a" containerID="89e5da178fa4be2fe40ec31859129846b5dc75e6b41ce9aa191a1ed7442e5dbf" exitCode=0 Dec 03 19:15:01 crc kubenswrapper[4758]: I1203 19:15:01.704879 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" event={"ID":"2aee032d-e890-4885-84ad-ce4dc9fbd20a","Type":"ContainerDied","Data":"89e5da178fa4be2fe40ec31859129846b5dc75e6b41ce9aa191a1ed7442e5dbf"} Dec 03 19:15:01 crc kubenswrapper[4758]: I1203 19:15:01.705157 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" event={"ID":"2aee032d-e890-4885-84ad-ce4dc9fbd20a","Type":"ContainerStarted","Data":"0b72ecb560a437a96967488dba89887a15f0ca006f861f70fc1c8a7f7ae7daf4"} Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.088570 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.089191 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.091082 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.156854 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.185319 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aee032d-e890-4885-84ad-ce4dc9fbd20a-config-volume\") pod \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.185490 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfbrm\" (UniqueName: \"kubernetes.io/projected/2aee032d-e890-4885-84ad-ce4dc9fbd20a-kube-api-access-vfbrm\") pod \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.185571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aee032d-e890-4885-84ad-ce4dc9fbd20a-secret-volume\") pod \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\" (UID: \"2aee032d-e890-4885-84ad-ce4dc9fbd20a\") " Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.186677 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aee032d-e890-4885-84ad-ce4dc9fbd20a-config-volume" (OuterVolumeSpecName: "config-volume") pod "2aee032d-e890-4885-84ad-ce4dc9fbd20a" (UID: "2aee032d-e890-4885-84ad-ce4dc9fbd20a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.201950 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee032d-e890-4885-84ad-ce4dc9fbd20a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2aee032d-e890-4885-84ad-ce4dc9fbd20a" (UID: "2aee032d-e890-4885-84ad-ce4dc9fbd20a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.202033 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aee032d-e890-4885-84ad-ce4dc9fbd20a-kube-api-access-vfbrm" (OuterVolumeSpecName: "kube-api-access-vfbrm") pod "2aee032d-e890-4885-84ad-ce4dc9fbd20a" (UID: "2aee032d-e890-4885-84ad-ce4dc9fbd20a"). InnerVolumeSpecName "kube-api-access-vfbrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.288218 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aee032d-e890-4885-84ad-ce4dc9fbd20a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.288254 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfbrm\" (UniqueName: \"kubernetes.io/projected/2aee032d-e890-4885-84ad-ce4dc9fbd20a-kube-api-access-vfbrm\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.288265 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aee032d-e890-4885-84ad-ce4dc9fbd20a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.728523 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.729348 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-4nlc6" event={"ID":"2aee032d-e890-4885-84ad-ce4dc9fbd20a","Type":"ContainerDied","Data":"0b72ecb560a437a96967488dba89887a15f0ca006f861f70fc1c8a7f7ae7daf4"} Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.729389 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b72ecb560a437a96967488dba89887a15f0ca006f861f70fc1c8a7f7ae7daf4" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.781531 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:15:03 crc kubenswrapper[4758]: I1203 19:15:03.833863 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lw69"] Dec 03 19:15:04 crc kubenswrapper[4758]: I1203 19:15:04.172109 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp"] Dec 03 19:15:04 crc kubenswrapper[4758]: I1203 19:15:04.182303 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-h6czp"] Dec 03 19:15:05 crc kubenswrapper[4758]: I1203 19:15:05.135000 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f" path="/var/lib/kubelet/pods/7a8af9a0-bf14-4d29-a5c8-f400a5dbd49f/volumes" Dec 03 19:15:05 crc kubenswrapper[4758]: I1203 19:15:05.745086 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4lw69" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="registry-server" containerID="cri-o://48403a04660aceeebbe46dfff8d83bce23964a19d045a2bff0fa51c678ea3d43" gracePeriod=2 Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.756509 4758 generic.go:334] "Generic (PLEG): container finished" podID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerID="48403a04660aceeebbe46dfff8d83bce23964a19d045a2bff0fa51c678ea3d43" exitCode=0 Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.756544 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lw69" event={"ID":"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4","Type":"ContainerDied","Data":"48403a04660aceeebbe46dfff8d83bce23964a19d045a2bff0fa51c678ea3d43"} Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.756821 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lw69" event={"ID":"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4","Type":"ContainerDied","Data":"74aa018f10bc939516e4b25e17b2d87610eed2c8a68125d5a54259f4d7d8194d"} Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.756836 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74aa018f10bc939516e4b25e17b2d87610eed2c8a68125d5a54259f4d7d8194d" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.759478 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.864938 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-utilities\") pod \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.865020 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc4x6\" (UniqueName: \"kubernetes.io/projected/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-kube-api-access-xc4x6\") pod \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.865063 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-catalog-content\") pod \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\" (UID: \"6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4\") " Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.866982 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-utilities" (OuterVolumeSpecName: "utilities") pod "6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" (UID: "6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.871198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-kube-api-access-xc4x6" (OuterVolumeSpecName: "kube-api-access-xc4x6") pod "6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" (UID: "6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4"). InnerVolumeSpecName "kube-api-access-xc4x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.890379 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" (UID: "6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.967881 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.968307 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc4x6\" (UniqueName: \"kubernetes.io/projected/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-kube-api-access-xc4x6\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:06 crc kubenswrapper[4758]: I1203 19:15:06.968325 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:07 crc kubenswrapper[4758]: I1203 19:15:07.765851 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lw69" Dec 03 19:15:07 crc kubenswrapper[4758]: I1203 19:15:07.794084 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lw69"] Dec 03 19:15:07 crc kubenswrapper[4758]: I1203 19:15:07.805951 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lw69"] Dec 03 19:15:09 crc kubenswrapper[4758]: I1203 19:15:09.129299 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" path="/var/lib/kubelet/pods/6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4/volumes" Dec 03 19:15:11 crc kubenswrapper[4758]: I1203 19:15:11.395777 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:15:11 crc kubenswrapper[4758]: I1203 19:15:11.396400 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:15:17 crc kubenswrapper[4758]: I1203 19:15:17.788221 4758 scope.go:117] "RemoveContainer" containerID="f8071202aca079ccd91262d1c98751537a40a01ad2c74d5b20373ff7c5bf4d86" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.928793 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2t4n2"] Dec 03 19:15:39 crc kubenswrapper[4758]: E1203 19:15:39.930102 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="registry-server" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.930122 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="registry-server" Dec 03 19:15:39 crc kubenswrapper[4758]: E1203 19:15:39.930170 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="extract-utilities" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.930180 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="extract-utilities" Dec 03 19:15:39 crc kubenswrapper[4758]: E1203 19:15:39.930204 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee032d-e890-4885-84ad-ce4dc9fbd20a" containerName="collect-profiles" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.930213 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee032d-e890-4885-84ad-ce4dc9fbd20a" containerName="collect-profiles" Dec 03 19:15:39 crc kubenswrapper[4758]: E1203 19:15:39.930229 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="extract-content" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.930235 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="extract-content" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.930464 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c3c7061-183b-4f19-a4f2-b9e56bfbd5f4" containerName="registry-server" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.930491 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee032d-e890-4885-84ad-ce4dc9fbd20a" containerName="collect-profiles" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.932496 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:39 crc kubenswrapper[4758]: I1203 19:15:39.956835 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2t4n2"] Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.113343 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjc9c\" (UniqueName: \"kubernetes.io/projected/f6e4b36e-447c-4884-b3e0-07fd593f4d80-kube-api-access-pjc9c\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.113749 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-utilities\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.114768 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-catalog-content\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.217381 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjc9c\" (UniqueName: \"kubernetes.io/projected/f6e4b36e-447c-4884-b3e0-07fd593f4d80-kube-api-access-pjc9c\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.217458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-utilities\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.217555 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-catalog-content\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.218099 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-utilities\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.218161 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-catalog-content\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.252220 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjc9c\" (UniqueName: \"kubernetes.io/projected/f6e4b36e-447c-4884-b3e0-07fd593f4d80-kube-api-access-pjc9c\") pod \"certified-operators-2t4n2\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.274241 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:40 crc kubenswrapper[4758]: I1203 19:15:40.894558 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2t4n2"] Dec 03 19:15:41 crc kubenswrapper[4758]: I1203 19:15:41.142515 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerStarted","Data":"b5944c58cdb43c0eef8d0c2d078d992d04669469b6bcc1e861763b703d44effd"} Dec 03 19:15:41 crc kubenswrapper[4758]: I1203 19:15:41.142974 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerStarted","Data":"8cedb6a503265b0f318e85ac216bf11b9714d6c9c0689a726b44fa2bf2f09020"} Dec 03 19:15:41 crc kubenswrapper[4758]: I1203 19:15:41.394384 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:15:41 crc kubenswrapper[4758]: I1203 19:15:41.394458 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:15:42 crc kubenswrapper[4758]: I1203 19:15:42.163757 4758 generic.go:334] "Generic (PLEG): container finished" podID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerID="b5944c58cdb43c0eef8d0c2d078d992d04669469b6bcc1e861763b703d44effd" exitCode=0 Dec 03 19:15:42 crc kubenswrapper[4758]: I1203 19:15:42.163834 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerDied","Data":"b5944c58cdb43c0eef8d0c2d078d992d04669469b6bcc1e861763b703d44effd"} Dec 03 19:15:42 crc kubenswrapper[4758]: I1203 19:15:42.170234 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:15:43 crc kubenswrapper[4758]: I1203 19:15:43.181616 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerStarted","Data":"eeef4ad7311d299e0f461dfdca37199406f12776b1e187e93ee7766d11dda411"} Dec 03 19:15:44 crc kubenswrapper[4758]: I1203 19:15:44.195867 4758 generic.go:334] "Generic (PLEG): container finished" podID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerID="eeef4ad7311d299e0f461dfdca37199406f12776b1e187e93ee7766d11dda411" exitCode=0 Dec 03 19:15:44 crc kubenswrapper[4758]: I1203 19:15:44.195938 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerDied","Data":"eeef4ad7311d299e0f461dfdca37199406f12776b1e187e93ee7766d11dda411"} Dec 03 19:15:45 crc kubenswrapper[4758]: I1203 19:15:45.220955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerStarted","Data":"6ccb226b53e654d735fb1d4020e8315bed84dc3dea658f12bb41fbc8bf959603"} Dec 03 19:15:45 crc kubenswrapper[4758]: I1203 19:15:45.252501 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2t4n2" podStartSLOduration=3.820110445 podStartE2EDuration="6.252481827s" podCreationTimestamp="2025-12-03 19:15:39 +0000 UTC" firstStartedPulling="2025-12-03 19:15:42.169894041 +0000 UTC m=+8397.371270902" lastFinishedPulling="2025-12-03 19:15:44.602265423 +0000 UTC m=+8399.803642284" observedRunningTime="2025-12-03 19:15:45.240830802 +0000 UTC m=+8400.442207663" watchObservedRunningTime="2025-12-03 19:15:45.252481827 +0000 UTC m=+8400.453858688" Dec 03 19:15:50 crc kubenswrapper[4758]: I1203 19:15:50.274745 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:50 crc kubenswrapper[4758]: I1203 19:15:50.275326 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:50 crc kubenswrapper[4758]: I1203 19:15:50.322578 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:51 crc kubenswrapper[4758]: I1203 19:15:51.328916 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:51 crc kubenswrapper[4758]: I1203 19:15:51.383363 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2t4n2"] Dec 03 19:15:53 crc kubenswrapper[4758]: I1203 19:15:53.296672 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2t4n2" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="registry-server" containerID="cri-o://6ccb226b53e654d735fb1d4020e8315bed84dc3dea658f12bb41fbc8bf959603" gracePeriod=2 Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.313819 4758 generic.go:334] "Generic (PLEG): container finished" podID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerID="6ccb226b53e654d735fb1d4020e8315bed84dc3dea658f12bb41fbc8bf959603" exitCode=0 Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.313904 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerDied","Data":"6ccb226b53e654d735fb1d4020e8315bed84dc3dea658f12bb41fbc8bf959603"} Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.738302 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.873580 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-catalog-content\") pod \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.874044 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-utilities\") pod \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.874210 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjc9c\" (UniqueName: \"kubernetes.io/projected/f6e4b36e-447c-4884-b3e0-07fd593f4d80-kube-api-access-pjc9c\") pod \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\" (UID: \"f6e4b36e-447c-4884-b3e0-07fd593f4d80\") " Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.875033 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-utilities" (OuterVolumeSpecName: "utilities") pod "f6e4b36e-447c-4884-b3e0-07fd593f4d80" (UID: "f6e4b36e-447c-4884-b3e0-07fd593f4d80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.888011 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6e4b36e-447c-4884-b3e0-07fd593f4d80-kube-api-access-pjc9c" (OuterVolumeSpecName: "kube-api-access-pjc9c") pod "f6e4b36e-447c-4884-b3e0-07fd593f4d80" (UID: "f6e4b36e-447c-4884-b3e0-07fd593f4d80"). InnerVolumeSpecName "kube-api-access-pjc9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.939172 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6e4b36e-447c-4884-b3e0-07fd593f4d80" (UID: "f6e4b36e-447c-4884-b3e0-07fd593f4d80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.976923 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.976959 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6e4b36e-447c-4884-b3e0-07fd593f4d80-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:54 crc kubenswrapper[4758]: I1203 19:15:54.976970 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjc9c\" (UniqueName: \"kubernetes.io/projected/f6e4b36e-447c-4884-b3e0-07fd593f4d80-kube-api-access-pjc9c\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.328104 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2t4n2" event={"ID":"f6e4b36e-447c-4884-b3e0-07fd593f4d80","Type":"ContainerDied","Data":"8cedb6a503265b0f318e85ac216bf11b9714d6c9c0689a726b44fa2bf2f09020"} Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.329522 4758 scope.go:117] "RemoveContainer" containerID="6ccb226b53e654d735fb1d4020e8315bed84dc3dea658f12bb41fbc8bf959603" Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.328181 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2t4n2" Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.361235 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2t4n2"] Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.369771 4758 scope.go:117] "RemoveContainer" containerID="eeef4ad7311d299e0f461dfdca37199406f12776b1e187e93ee7766d11dda411" Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.372579 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2t4n2"] Dec 03 19:15:55 crc kubenswrapper[4758]: I1203 19:15:55.391992 4758 scope.go:117] "RemoveContainer" containerID="b5944c58cdb43c0eef8d0c2d078d992d04669469b6bcc1e861763b703d44effd" Dec 03 19:15:57 crc kubenswrapper[4758]: I1203 19:15:57.126896 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" path="/var/lib/kubelet/pods/f6e4b36e-447c-4884-b3e0-07fd593f4d80/volumes" Dec 03 19:15:57 crc kubenswrapper[4758]: I1203 19:15:57.381608 4758 generic.go:334] "Generic (PLEG): container finished" podID="0594107d-92c8-4f63-9d64-5c362703c627" containerID="2da5a80861307b8ceba5d5d38d88f539ac7c3d0455664bc129e9295036dcc5fb" exitCode=0 Dec 03 19:15:57 crc kubenswrapper[4758]: I1203 19:15:57.381660 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" event={"ID":"0594107d-92c8-4f63-9d64-5c362703c627","Type":"ContainerDied","Data":"2da5a80861307b8ceba5d5d38d88f539ac7c3d0455664bc129e9295036dcc5fb"} Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.868336 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.971960 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-inventory\") pod \"0594107d-92c8-4f63-9d64-5c362703c627\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.972021 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8t7v\" (UniqueName: \"kubernetes.io/projected/0594107d-92c8-4f63-9d64-5c362703c627-kube-api-access-q8t7v\") pod \"0594107d-92c8-4f63-9d64-5c362703c627\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.972091 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-combined-ca-bundle\") pod \"0594107d-92c8-4f63-9d64-5c362703c627\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.972140 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ssh-key\") pod \"0594107d-92c8-4f63-9d64-5c362703c627\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.972193 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-secret-0\") pod \"0594107d-92c8-4f63-9d64-5c362703c627\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.972298 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ceph\") pod \"0594107d-92c8-4f63-9d64-5c362703c627\" (UID: \"0594107d-92c8-4f63-9d64-5c362703c627\") " Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.978634 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ceph" (OuterVolumeSpecName: "ceph") pod "0594107d-92c8-4f63-9d64-5c362703c627" (UID: "0594107d-92c8-4f63-9d64-5c362703c627"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.979514 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0594107d-92c8-4f63-9d64-5c362703c627-kube-api-access-q8t7v" (OuterVolumeSpecName: "kube-api-access-q8t7v") pod "0594107d-92c8-4f63-9d64-5c362703c627" (UID: "0594107d-92c8-4f63-9d64-5c362703c627"). InnerVolumeSpecName "kube-api-access-q8t7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:15:58 crc kubenswrapper[4758]: I1203 19:15:58.980333 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0594107d-92c8-4f63-9d64-5c362703c627" (UID: "0594107d-92c8-4f63-9d64-5c362703c627"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.017560 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "0594107d-92c8-4f63-9d64-5c362703c627" (UID: "0594107d-92c8-4f63-9d64-5c362703c627"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.021149 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0594107d-92c8-4f63-9d64-5c362703c627" (UID: "0594107d-92c8-4f63-9d64-5c362703c627"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.053652 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-inventory" (OuterVolumeSpecName: "inventory") pod "0594107d-92c8-4f63-9d64-5c362703c627" (UID: "0594107d-92c8-4f63-9d64-5c362703c627"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.074954 4758 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.074986 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.074996 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.075008 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8t7v\" (UniqueName: \"kubernetes.io/projected/0594107d-92c8-4f63-9d64-5c362703c627-kube-api-access-q8t7v\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.075018 4758 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.075027 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0594107d-92c8-4f63-9d64-5c362703c627-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.415241 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" event={"ID":"0594107d-92c8-4f63-9d64-5c362703c627","Type":"ContainerDied","Data":"85232d39884e3013295752f1555aa31ee9696180401484d18cbce827b7587073"} Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.415294 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85232d39884e3013295752f1555aa31ee9696180401484d18cbce827b7587073" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.415400 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-wkqhj" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.527848 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-bsb6m"] Dec 03 19:15:59 crc kubenswrapper[4758]: E1203 19:15:59.528422 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0594107d-92c8-4f63-9d64-5c362703c627" containerName="libvirt-openstack-openstack-cell1" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.528446 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0594107d-92c8-4f63-9d64-5c362703c627" containerName="libvirt-openstack-openstack-cell1" Dec 03 19:15:59 crc kubenswrapper[4758]: E1203 19:15:59.528470 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="extract-utilities" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.528483 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="extract-utilities" Dec 03 19:15:59 crc kubenswrapper[4758]: E1203 19:15:59.528521 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="extract-content" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.528533 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="extract-content" Dec 03 19:15:59 crc kubenswrapper[4758]: E1203 19:15:59.528559 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="registry-server" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.528569 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="registry-server" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.528837 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e4b36e-447c-4884-b3e0-07fd593f4d80" containerName="registry-server" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.528864 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0594107d-92c8-4f63-9d64-5c362703c627" containerName="libvirt-openstack-openstack-cell1" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.530333 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533274 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533402 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533466 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533676 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533801 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533854 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.533944 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.542295 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-bsb6m"] Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.586433 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.586868 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ceph\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.586904 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.586921 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587116 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587175 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587196 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-inventory\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587274 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfgtb\" (UniqueName: \"kubernetes.io/projected/d6d965b5-a0a8-4ba8-a25c-623612615e02-kube-api-access-qfgtb\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587316 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587338 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.587364 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.689403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ceph\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.689778 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.689910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690082 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690234 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-inventory\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfgtb\" (UniqueName: \"kubernetes.io/projected/d6d965b5-a0a8-4ba8-a25c-623612615e02-kube-api-access-qfgtb\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690682 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690855 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.690998 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.691184 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.692259 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.692434 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.693915 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-inventory\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.693978 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.694046 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.694831 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ceph\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.696095 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.696367 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.696968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.703596 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.707110 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfgtb\" (UniqueName: \"kubernetes.io/projected/d6d965b5-a0a8-4ba8-a25c-623612615e02-kube-api-access-qfgtb\") pod \"nova-cell1-openstack-openstack-cell1-bsb6m\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:15:59 crc kubenswrapper[4758]: I1203 19:15:59.867088 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:16:00 crc kubenswrapper[4758]: I1203 19:16:00.408376 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-bsb6m"] Dec 03 19:16:00 crc kubenswrapper[4758]: I1203 19:16:00.428758 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" event={"ID":"d6d965b5-a0a8-4ba8-a25c-623612615e02","Type":"ContainerStarted","Data":"bbfea1da64af74161dfe567cd6ae850f0be17aadf99cceea2738d1a981ba5906"} Dec 03 19:16:03 crc kubenswrapper[4758]: I1203 19:16:03.460014 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" event={"ID":"d6d965b5-a0a8-4ba8-a25c-623612615e02","Type":"ContainerStarted","Data":"38d1fc52e121a0acf2c62c61f5748650ce2fcc55aeeb671b0d1021fbb93c813c"} Dec 03 19:16:03 crc kubenswrapper[4758]: I1203 19:16:03.492575 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" podStartSLOduration=2.462519113 podStartE2EDuration="4.492556691s" podCreationTimestamp="2025-12-03 19:15:59 +0000 UTC" firstStartedPulling="2025-12-03 19:16:00.414492266 +0000 UTC m=+8415.615869127" lastFinishedPulling="2025-12-03 19:16:02.444529844 +0000 UTC m=+8417.645906705" observedRunningTime="2025-12-03 19:16:03.4825585 +0000 UTC m=+8418.683935361" watchObservedRunningTime="2025-12-03 19:16:03.492556691 +0000 UTC m=+8418.693933542" Dec 03 19:16:11 crc kubenswrapper[4758]: I1203 19:16:11.395598 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:16:11 crc kubenswrapper[4758]: I1203 19:16:11.396364 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:16:11 crc kubenswrapper[4758]: I1203 19:16:11.396802 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:16:11 crc kubenswrapper[4758]: I1203 19:16:11.397864 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"435e139b7660730a3057119bb44eca7c2106ea848a8a6724b9ce30a20679541a"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:16:11 crc kubenswrapper[4758]: I1203 19:16:11.397925 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://435e139b7660730a3057119bb44eca7c2106ea848a8a6724b9ce30a20679541a" gracePeriod=600 Dec 03 19:16:12 crc kubenswrapper[4758]: I1203 19:16:12.600132 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="435e139b7660730a3057119bb44eca7c2106ea848a8a6724b9ce30a20679541a" exitCode=0 Dec 03 19:16:12 crc kubenswrapper[4758]: I1203 19:16:12.600442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"435e139b7660730a3057119bb44eca7c2106ea848a8a6724b9ce30a20679541a"} Dec 03 19:16:12 crc kubenswrapper[4758]: I1203 19:16:12.600470 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4"} Dec 03 19:16:12 crc kubenswrapper[4758]: I1203 19:16:12.600486 4758 scope.go:117] "RemoveContainer" containerID="85789b78fba326ff47d45838970624af5c5e870df118bf840ceab85f288055e1" Dec 03 19:18:41 crc kubenswrapper[4758]: I1203 19:18:41.394764 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:18:41 crc kubenswrapper[4758]: I1203 19:18:41.395269 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:18:44 crc kubenswrapper[4758]: I1203 19:18:44.140021 4758 generic.go:334] "Generic (PLEG): container finished" podID="d6d965b5-a0a8-4ba8-a25c-623612615e02" containerID="38d1fc52e121a0acf2c62c61f5748650ce2fcc55aeeb671b0d1021fbb93c813c" exitCode=0 Dec 03 19:18:44 crc kubenswrapper[4758]: I1203 19:18:44.140117 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" event={"ID":"d6d965b5-a0a8-4ba8-a25c-623612615e02","Type":"ContainerDied","Data":"38d1fc52e121a0acf2c62c61f5748650ce2fcc55aeeb671b0d1021fbb93c813c"} Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.634565 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696339 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-1\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696390 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfgtb\" (UniqueName: \"kubernetes.io/projected/d6d965b5-a0a8-4ba8-a25c-623612615e02-kube-api-access-qfgtb\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-combined-ca-bundle\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696433 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-1\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696496 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-0\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696544 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-0\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696655 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-1\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696721 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-0\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696825 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ssh-key\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696879 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ceph\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.696905 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-inventory\") pod \"d6d965b5-a0a8-4ba8-a25c-623612615e02\" (UID: \"d6d965b5-a0a8-4ba8-a25c-623612615e02\") " Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.702927 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ceph" (OuterVolumeSpecName: "ceph") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.704971 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.719830 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d965b5-a0a8-4ba8-a25c-623612615e02-kube-api-access-qfgtb" (OuterVolumeSpecName: "kube-api-access-qfgtb") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "kube-api-access-qfgtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.728518 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.733651 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-inventory" (OuterVolumeSpecName: "inventory") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.735200 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.738333 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.739376 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.740363 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.740910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.743058 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6d965b5-a0a8-4ba8-a25c-623612615e02" (UID: "d6d965b5-a0a8-4ba8-a25c-623612615e02"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799242 4758 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799502 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799567 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799625 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799696 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799764 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799826 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfgtb\" (UniqueName: \"kubernetes.io/projected/d6d965b5-a0a8-4ba8-a25c-623612615e02-kube-api-access-qfgtb\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799883 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.799975 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.800047 4758 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:45 crc kubenswrapper[4758]: I1203 19:18:45.800110 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6d965b5-a0a8-4ba8-a25c-623612615e02-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.161576 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" event={"ID":"d6d965b5-a0a8-4ba8-a25c-623612615e02","Type":"ContainerDied","Data":"bbfea1da64af74161dfe567cd6ae850f0be17aadf99cceea2738d1a981ba5906"} Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.161859 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbfea1da64af74161dfe567cd6ae850f0be17aadf99cceea2738d1a981ba5906" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.161623 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bsb6m" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.260765 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-pbnq4"] Dec 03 19:18:46 crc kubenswrapper[4758]: E1203 19:18:46.261411 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d965b5-a0a8-4ba8-a25c-623612615e02" containerName="nova-cell1-openstack-openstack-cell1" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.261438 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d965b5-a0a8-4ba8-a25c-623612615e02" containerName="nova-cell1-openstack-openstack-cell1" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.261847 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d965b5-a0a8-4ba8-a25c-623612615e02" containerName="nova-cell1-openstack-openstack-cell1" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.262853 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.264962 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.265277 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.265471 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.265482 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.265628 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.270988 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-pbnq4"] Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.310831 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-inventory\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.310959 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.310998 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m5r8\" (UniqueName: \"kubernetes.io/projected/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-kube-api-access-6m5r8\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.311031 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.311357 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.311436 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceph\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.311779 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.311841 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ssh-key\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413151 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413283 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceph\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ssh-key\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413500 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-inventory\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413535 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.413568 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m5r8\" (UniqueName: \"kubernetes.io/projected/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-kube-api-access-6m5r8\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.418317 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.418815 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.419186 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ssh-key\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.419802 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-inventory\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.420430 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.421145 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceph\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.425142 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.430485 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m5r8\" (UniqueName: \"kubernetes.io/projected/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-kube-api-access-6m5r8\") pod \"telemetry-openstack-openstack-cell1-pbnq4\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:46 crc kubenswrapper[4758]: I1203 19:18:46.589546 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:18:47 crc kubenswrapper[4758]: I1203 19:18:47.115178 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-pbnq4"] Dec 03 19:18:47 crc kubenswrapper[4758]: I1203 19:18:47.172204 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" event={"ID":"fa3c885d-c09e-4fdf-ae60-61bbfc92df32","Type":"ContainerStarted","Data":"5add6d171ebf5d3c32e8e1f6ec38e3b3e6c350819c806387b26ca8aa584c556a"} Dec 03 19:18:48 crc kubenswrapper[4758]: I1203 19:18:48.187240 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" event={"ID":"fa3c885d-c09e-4fdf-ae60-61bbfc92df32","Type":"ContainerStarted","Data":"d74506e58a1117a82ebb8b723ca1d71a351b07cf796eb8f7a321673f713ec0d8"} Dec 03 19:18:48 crc kubenswrapper[4758]: I1203 19:18:48.217778 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" podStartSLOduration=2.024481372 podStartE2EDuration="2.217756105s" podCreationTimestamp="2025-12-03 19:18:46 +0000 UTC" firstStartedPulling="2025-12-03 19:18:47.126770258 +0000 UTC m=+8582.328147119" lastFinishedPulling="2025-12-03 19:18:47.320044981 +0000 UTC m=+8582.521421852" observedRunningTime="2025-12-03 19:18:48.206527914 +0000 UTC m=+8583.407904805" watchObservedRunningTime="2025-12-03 19:18:48.217756105 +0000 UTC m=+8583.419132966" Dec 03 19:19:11 crc kubenswrapper[4758]: I1203 19:19:11.394779 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:19:11 crc kubenswrapper[4758]: I1203 19:19:11.395314 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.394219 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.394810 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.394865 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.395488 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.395539 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" gracePeriod=600 Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.745575 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" exitCode=0 Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.745644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4"} Dec 03 19:19:41 crc kubenswrapper[4758]: I1203 19:19:41.746045 4758 scope.go:117] "RemoveContainer" containerID="435e139b7660730a3057119bb44eca7c2106ea848a8a6724b9ce30a20679541a" Dec 03 19:19:42 crc kubenswrapper[4758]: E1203 19:19:42.029360 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:19:42 crc kubenswrapper[4758]: I1203 19:19:42.756970 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:19:42 crc kubenswrapper[4758]: E1203 19:19:42.758303 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:19:56 crc kubenswrapper[4758]: I1203 19:19:56.115213 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:19:56 crc kubenswrapper[4758]: E1203 19:19:56.116246 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:20:07 crc kubenswrapper[4758]: I1203 19:20:07.114735 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:20:07 crc kubenswrapper[4758]: E1203 19:20:07.115606 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:20:18 crc kubenswrapper[4758]: I1203 19:20:18.115104 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:20:18 crc kubenswrapper[4758]: E1203 19:20:18.115930 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:20:30 crc kubenswrapper[4758]: I1203 19:20:30.113971 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:20:30 crc kubenswrapper[4758]: E1203 19:20:30.115000 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:20:44 crc kubenswrapper[4758]: I1203 19:20:44.114693 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:20:44 crc kubenswrapper[4758]: E1203 19:20:44.115567 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:20:56 crc kubenswrapper[4758]: I1203 19:20:56.115163 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:20:56 crc kubenswrapper[4758]: E1203 19:20:56.116070 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:21:11 crc kubenswrapper[4758]: I1203 19:21:11.114667 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:21:11 crc kubenswrapper[4758]: E1203 19:21:11.115489 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:21:18 crc kubenswrapper[4758]: I1203 19:21:18.006364 4758 scope.go:117] "RemoveContainer" containerID="a4ec28693c7b0a2ec9314b6e8c49c08b3316377de8df158427860055edb33b5f" Dec 03 19:21:18 crc kubenswrapper[4758]: I1203 19:21:18.046874 4758 scope.go:117] "RemoveContainer" containerID="069b497884c33d9d7fc2ab60d56c474577bde955220fcbc2a45b81a43563f0fa" Dec 03 19:21:18 crc kubenswrapper[4758]: I1203 19:21:18.087957 4758 scope.go:117] "RemoveContainer" containerID="48403a04660aceeebbe46dfff8d83bce23964a19d045a2bff0fa51c678ea3d43" Dec 03 19:21:25 crc kubenswrapper[4758]: I1203 19:21:25.122856 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:21:25 crc kubenswrapper[4758]: E1203 19:21:25.123856 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:21:37 crc kubenswrapper[4758]: I1203 19:21:37.114385 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:21:37 crc kubenswrapper[4758]: E1203 19:21:37.115355 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:21:50 crc kubenswrapper[4758]: I1203 19:21:50.113897 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:21:50 crc kubenswrapper[4758]: E1203 19:21:50.115814 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:22:03 crc kubenswrapper[4758]: I1203 19:22:03.115059 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:22:03 crc kubenswrapper[4758]: E1203 19:22:03.115979 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:22:18 crc kubenswrapper[4758]: I1203 19:22:18.115006 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:22:18 crc kubenswrapper[4758]: E1203 19:22:18.117522 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:22:32 crc kubenswrapper[4758]: I1203 19:22:32.113907 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:22:32 crc kubenswrapper[4758]: E1203 19:22:32.114796 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:22:44 crc kubenswrapper[4758]: I1203 19:22:44.581644 4758 generic.go:334] "Generic (PLEG): container finished" podID="fa3c885d-c09e-4fdf-ae60-61bbfc92df32" containerID="d74506e58a1117a82ebb8b723ca1d71a351b07cf796eb8f7a321673f713ec0d8" exitCode=0 Dec 03 19:22:44 crc kubenswrapper[4758]: I1203 19:22:44.581739 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" event={"ID":"fa3c885d-c09e-4fdf-ae60-61bbfc92df32","Type":"ContainerDied","Data":"d74506e58a1117a82ebb8b723ca1d71a351b07cf796eb8f7a321673f713ec0d8"} Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.063341 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.104802 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-0\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.104877 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ssh-key\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.105018 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-2\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.105242 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-telemetry-combined-ca-bundle\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.105274 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceph\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.105433 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-inventory\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.106166 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m5r8\" (UniqueName: \"kubernetes.io/projected/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-kube-api-access-6m5r8\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.106233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-1\") pod \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\" (UID: \"fa3c885d-c09e-4fdf-ae60-61bbfc92df32\") " Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.111113 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceph" (OuterVolumeSpecName: "ceph") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.111940 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.112445 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-kube-api-access-6m5r8" (OuterVolumeSpecName: "kube-api-access-6m5r8") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "kube-api-access-6m5r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.114601 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:22:46 crc kubenswrapper[4758]: E1203 19:22:46.115093 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.133907 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.136275 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.136615 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.147268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.147923 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-inventory" (OuterVolumeSpecName: "inventory") pod "fa3c885d-c09e-4fdf-ae60-61bbfc92df32" (UID: "fa3c885d-c09e-4fdf-ae60-61bbfc92df32"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208550 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208607 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208622 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208634 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208645 4758 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208657 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208704 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.208714 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m5r8\" (UniqueName: \"kubernetes.io/projected/fa3c885d-c09e-4fdf-ae60-61bbfc92df32-kube-api-access-6m5r8\") on node \"crc\" DevicePath \"\"" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.613473 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" event={"ID":"fa3c885d-c09e-4fdf-ae60-61bbfc92df32","Type":"ContainerDied","Data":"5add6d171ebf5d3c32e8e1f6ec38e3b3e6c350819c806387b26ca8aa584c556a"} Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.613876 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5add6d171ebf5d3c32e8e1f6ec38e3b3e6c350819c806387b26ca8aa584c556a" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.613552 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-pbnq4" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.700296 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-tzxnl"] Dec 03 19:22:46 crc kubenswrapper[4758]: E1203 19:22:46.700843 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3c885d-c09e-4fdf-ae60-61bbfc92df32" containerName="telemetry-openstack-openstack-cell1" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.700863 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3c885d-c09e-4fdf-ae60-61bbfc92df32" containerName="telemetry-openstack-openstack-cell1" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.701131 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa3c885d-c09e-4fdf-ae60-61bbfc92df32" containerName="telemetry-openstack-openstack-cell1" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.702007 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.705040 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.705313 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.705587 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.706771 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.707900 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.718319 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-tzxnl"] Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.819873 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.819941 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.820022 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.820078 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-955t8\" (UniqueName: \"kubernetes.io/projected/80659ad8-2251-46bb-853d-18ac040f6d84-kube-api-access-955t8\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.820119 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.820206 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.922272 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.922324 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.922370 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.922441 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.922504 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-955t8\" (UniqueName: \"kubernetes.io/projected/80659ad8-2251-46bb-853d-18ac040f6d84-kube-api-access-955t8\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.922567 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.927519 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.927781 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.927807 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.927792 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.928439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:46 crc kubenswrapper[4758]: I1203 19:22:46.943934 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-955t8\" (UniqueName: \"kubernetes.io/projected/80659ad8-2251-46bb-853d-18ac040f6d84-kube-api-access-955t8\") pod \"neutron-sriov-openstack-openstack-cell1-tzxnl\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:47 crc kubenswrapper[4758]: I1203 19:22:47.022875 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:22:47 crc kubenswrapper[4758]: I1203 19:22:47.582013 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-tzxnl"] Dec 03 19:22:47 crc kubenswrapper[4758]: W1203 19:22:47.597804 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80659ad8_2251_46bb_853d_18ac040f6d84.slice/crio-3e4997adfe3b28b149baf79bcedf9ee13b29448380e3c12ab9ad11e70c388220 WatchSource:0}: Error finding container 3e4997adfe3b28b149baf79bcedf9ee13b29448380e3c12ab9ad11e70c388220: Status 404 returned error can't find the container with id 3e4997adfe3b28b149baf79bcedf9ee13b29448380e3c12ab9ad11e70c388220 Dec 03 19:22:47 crc kubenswrapper[4758]: I1203 19:22:47.602861 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:22:47 crc kubenswrapper[4758]: I1203 19:22:47.627708 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" event={"ID":"80659ad8-2251-46bb-853d-18ac040f6d84","Type":"ContainerStarted","Data":"3e4997adfe3b28b149baf79bcedf9ee13b29448380e3c12ab9ad11e70c388220"} Dec 03 19:22:48 crc kubenswrapper[4758]: I1203 19:22:48.639725 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" event={"ID":"80659ad8-2251-46bb-853d-18ac040f6d84","Type":"ContainerStarted","Data":"a6a76c46b9e657ed9e07ad1f5086d383825228ec035fee46acbefcd1b564d694"} Dec 03 19:22:48 crc kubenswrapper[4758]: I1203 19:22:48.661594 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" podStartSLOduration=2.376584629 podStartE2EDuration="2.661576132s" podCreationTimestamp="2025-12-03 19:22:46 +0000 UTC" firstStartedPulling="2025-12-03 19:22:47.602622474 +0000 UTC m=+8822.803999335" lastFinishedPulling="2025-12-03 19:22:47.887613977 +0000 UTC m=+8823.088990838" observedRunningTime="2025-12-03 19:22:48.656515027 +0000 UTC m=+8823.857891888" watchObservedRunningTime="2025-12-03 19:22:48.661576132 +0000 UTC m=+8823.862952993" Dec 03 19:22:58 crc kubenswrapper[4758]: I1203 19:22:58.114247 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:22:58 crc kubenswrapper[4758]: E1203 19:22:58.115171 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:23:11 crc kubenswrapper[4758]: I1203 19:23:11.115118 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:23:11 crc kubenswrapper[4758]: E1203 19:23:11.115919 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:23:23 crc kubenswrapper[4758]: I1203 19:23:23.115813 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:23:23 crc kubenswrapper[4758]: E1203 19:23:23.116835 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:23:34 crc kubenswrapper[4758]: I1203 19:23:34.114751 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:23:34 crc kubenswrapper[4758]: E1203 19:23:34.116796 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:23:49 crc kubenswrapper[4758]: I1203 19:23:49.115421 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:23:49 crc kubenswrapper[4758]: E1203 19:23:49.116465 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:24:02 crc kubenswrapper[4758]: I1203 19:24:02.115090 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:24:02 crc kubenswrapper[4758]: E1203 19:24:02.116203 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:24:17 crc kubenswrapper[4758]: I1203 19:24:17.114256 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:24:17 crc kubenswrapper[4758]: E1203 19:24:17.115267 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:24:32 crc kubenswrapper[4758]: I1203 19:24:32.115672 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:24:32 crc kubenswrapper[4758]: E1203 19:24:32.117034 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:24:43 crc kubenswrapper[4758]: I1203 19:24:43.114940 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:24:43 crc kubenswrapper[4758]: I1203 19:24:43.930570 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"0ee4ad7fd5bf09c0c2f45187fe086f2bafb84a762e68492bc44a65b75f10a229"} Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.688444 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qxbf4"] Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.691959 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.701039 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxbf4"] Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.794548 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl2cc\" (UniqueName: \"kubernetes.io/projected/636fc5d6-ebf8-492a-9b98-27eb1950a693-kube-api-access-wl2cc\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.795079 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-catalog-content\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.795124 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-utilities\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.898105 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl2cc\" (UniqueName: \"kubernetes.io/projected/636fc5d6-ebf8-492a-9b98-27eb1950a693-kube-api-access-wl2cc\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.898228 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-catalog-content\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.898299 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-utilities\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.898998 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-utilities\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.899046 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-catalog-content\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:52 crc kubenswrapper[4758]: I1203 19:24:52.932354 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl2cc\" (UniqueName: \"kubernetes.io/projected/636fc5d6-ebf8-492a-9b98-27eb1950a693-kube-api-access-wl2cc\") pod \"community-operators-qxbf4\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:53 crc kubenswrapper[4758]: I1203 19:24:53.025492 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:24:53 crc kubenswrapper[4758]: I1203 19:24:53.648938 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxbf4"] Dec 03 19:24:53 crc kubenswrapper[4758]: W1203 19:24:53.652217 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod636fc5d6_ebf8_492a_9b98_27eb1950a693.slice/crio-f8b647253f989a861faca363155b364129ad32947c3390a2da044865de9ee079 WatchSource:0}: Error finding container f8b647253f989a861faca363155b364129ad32947c3390a2da044865de9ee079: Status 404 returned error can't find the container with id f8b647253f989a861faca363155b364129ad32947c3390a2da044865de9ee079 Dec 03 19:24:54 crc kubenswrapper[4758]: I1203 19:24:54.040266 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerStarted","Data":"381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5"} Dec 03 19:24:54 crc kubenswrapper[4758]: I1203 19:24:54.040595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerStarted","Data":"f8b647253f989a861faca363155b364129ad32947c3390a2da044865de9ee079"} Dec 03 19:24:55 crc kubenswrapper[4758]: I1203 19:24:55.053210 4758 generic.go:334] "Generic (PLEG): container finished" podID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerID="381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5" exitCode=0 Dec 03 19:24:55 crc kubenswrapper[4758]: I1203 19:24:55.053313 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerDied","Data":"381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5"} Dec 03 19:24:56 crc kubenswrapper[4758]: I1203 19:24:56.076663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerStarted","Data":"7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452"} Dec 03 19:24:57 crc kubenswrapper[4758]: I1203 19:24:57.088220 4758 generic.go:334] "Generic (PLEG): container finished" podID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerID="7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452" exitCode=0 Dec 03 19:24:57 crc kubenswrapper[4758]: I1203 19:24:57.088337 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerDied","Data":"7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452"} Dec 03 19:24:58 crc kubenswrapper[4758]: I1203 19:24:58.101139 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerStarted","Data":"1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d"} Dec 03 19:24:58 crc kubenswrapper[4758]: I1203 19:24:58.125527 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qxbf4" podStartSLOduration=3.619414526 podStartE2EDuration="6.125497473s" podCreationTimestamp="2025-12-03 19:24:52 +0000 UTC" firstStartedPulling="2025-12-03 19:24:55.056144821 +0000 UTC m=+8950.257521712" lastFinishedPulling="2025-12-03 19:24:57.562227798 +0000 UTC m=+8952.763604659" observedRunningTime="2025-12-03 19:24:58.117108338 +0000 UTC m=+8953.318485199" watchObservedRunningTime="2025-12-03 19:24:58.125497473 +0000 UTC m=+8953.326874334" Dec 03 19:25:03 crc kubenswrapper[4758]: I1203 19:25:03.026058 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:25:03 crc kubenswrapper[4758]: I1203 19:25:03.026580 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:25:03 crc kubenswrapper[4758]: I1203 19:25:03.088278 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:25:03 crc kubenswrapper[4758]: I1203 19:25:03.195260 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:25:03 crc kubenswrapper[4758]: I1203 19:25:03.322844 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxbf4"] Dec 03 19:25:05 crc kubenswrapper[4758]: I1203 19:25:05.165115 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qxbf4" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="registry-server" containerID="cri-o://1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d" gracePeriod=2 Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.148732 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.176868 4758 generic.go:334] "Generic (PLEG): container finished" podID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerID="1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d" exitCode=0 Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.176925 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerDied","Data":"1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d"} Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.176937 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxbf4" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.176974 4758 scope.go:117] "RemoveContainer" containerID="1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.176961 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxbf4" event={"ID":"636fc5d6-ebf8-492a-9b98-27eb1950a693","Type":"ContainerDied","Data":"f8b647253f989a861faca363155b364129ad32947c3390a2da044865de9ee079"} Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.208111 4758 scope.go:117] "RemoveContainer" containerID="7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.243906 4758 scope.go:117] "RemoveContainer" containerID="381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.278355 4758 scope.go:117] "RemoveContainer" containerID="1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d" Dec 03 19:25:06 crc kubenswrapper[4758]: E1203 19:25:06.278989 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d\": container with ID starting with 1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d not found: ID does not exist" containerID="1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.279049 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d"} err="failed to get container status \"1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d\": rpc error: code = NotFound desc = could not find container \"1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d\": container with ID starting with 1fb61f14d42c7a7a3b44e979fe1dea876ca855ab207d8c6b55371cb1bf46970d not found: ID does not exist" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.279086 4758 scope.go:117] "RemoveContainer" containerID="7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452" Dec 03 19:25:06 crc kubenswrapper[4758]: E1203 19:25:06.279437 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452\": container with ID starting with 7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452 not found: ID does not exist" containerID="7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.279479 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452"} err="failed to get container status \"7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452\": rpc error: code = NotFound desc = could not find container \"7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452\": container with ID starting with 7d56706fea37642d1cd17d6c0ced5a83455e4573f99bf5abcbc0d7dcf9a23452 not found: ID does not exist" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.279509 4758 scope.go:117] "RemoveContainer" containerID="381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5" Dec 03 19:25:06 crc kubenswrapper[4758]: E1203 19:25:06.279808 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5\": container with ID starting with 381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5 not found: ID does not exist" containerID="381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.279847 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5"} err="failed to get container status \"381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5\": rpc error: code = NotFound desc = could not find container \"381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5\": container with ID starting with 381a36adf0cfd84325f23b1308fc24c7191f90dbfef7fc28f986660f433b8fb5 not found: ID does not exist" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.313441 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-utilities\") pod \"636fc5d6-ebf8-492a-9b98-27eb1950a693\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.313522 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-catalog-content\") pod \"636fc5d6-ebf8-492a-9b98-27eb1950a693\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.313829 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl2cc\" (UniqueName: \"kubernetes.io/projected/636fc5d6-ebf8-492a-9b98-27eb1950a693-kube-api-access-wl2cc\") pod \"636fc5d6-ebf8-492a-9b98-27eb1950a693\" (UID: \"636fc5d6-ebf8-492a-9b98-27eb1950a693\") " Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.316300 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-utilities" (OuterVolumeSpecName: "utilities") pod "636fc5d6-ebf8-492a-9b98-27eb1950a693" (UID: "636fc5d6-ebf8-492a-9b98-27eb1950a693"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.320072 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636fc5d6-ebf8-492a-9b98-27eb1950a693-kube-api-access-wl2cc" (OuterVolumeSpecName: "kube-api-access-wl2cc") pod "636fc5d6-ebf8-492a-9b98-27eb1950a693" (UID: "636fc5d6-ebf8-492a-9b98-27eb1950a693"). InnerVolumeSpecName "kube-api-access-wl2cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.369154 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "636fc5d6-ebf8-492a-9b98-27eb1950a693" (UID: "636fc5d6-ebf8-492a-9b98-27eb1950a693"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.416135 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.416184 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl2cc\" (UniqueName: \"kubernetes.io/projected/636fc5d6-ebf8-492a-9b98-27eb1950a693-kube-api-access-wl2cc\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.416200 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636fc5d6-ebf8-492a-9b98-27eb1950a693-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.527080 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxbf4"] Dec 03 19:25:06 crc kubenswrapper[4758]: I1203 19:25:06.536744 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qxbf4"] Dec 03 19:25:07 crc kubenswrapper[4758]: I1203 19:25:07.127603 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" path="/var/lib/kubelet/pods/636fc5d6-ebf8-492a-9b98-27eb1950a693/volumes" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.021985 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xs5p7"] Dec 03 19:25:14 crc kubenswrapper[4758]: E1203 19:25:14.024603 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="extract-content" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.024629 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="extract-content" Dec 03 19:25:14 crc kubenswrapper[4758]: E1203 19:25:14.024657 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="registry-server" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.024666 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="registry-server" Dec 03 19:25:14 crc kubenswrapper[4758]: E1203 19:25:14.024695 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="extract-utilities" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.024709 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="extract-utilities" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.025079 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="636fc5d6-ebf8-492a-9b98-27eb1950a693" containerName="registry-server" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.028457 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.043522 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xs5p7"] Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.182355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-catalog-content\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.182451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdsrb\" (UniqueName: \"kubernetes.io/projected/e0328a71-b1d5-412e-bd4c-776ff76c1157-kube-api-access-kdsrb\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.182633 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-utilities\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.213104 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2rqqj"] Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.216456 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.228514 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rqqj"] Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.284543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-catalog-content\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.284621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdsrb\" (UniqueName: \"kubernetes.io/projected/e0328a71-b1d5-412e-bd4c-776ff76c1157-kube-api-access-kdsrb\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.284663 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-utilities\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.285107 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-catalog-content\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.285409 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-utilities\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.306057 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdsrb\" (UniqueName: \"kubernetes.io/projected/e0328a71-b1d5-412e-bd4c-776ff76c1157-kube-api-access-kdsrb\") pod \"redhat-marketplace-xs5p7\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.364278 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.387451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-utilities\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.387533 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-catalog-content\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.387786 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hw77\" (UniqueName: \"kubernetes.io/projected/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-kube-api-access-9hw77\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.490409 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-utilities\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.490480 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-catalog-content\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.490540 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hw77\" (UniqueName: \"kubernetes.io/projected/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-kube-api-access-9hw77\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.491025 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-utilities\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.491318 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-catalog-content\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.514654 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hw77\" (UniqueName: \"kubernetes.io/projected/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-kube-api-access-9hw77\") pod \"redhat-operators-2rqqj\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.539946 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:14 crc kubenswrapper[4758]: I1203 19:25:14.957492 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xs5p7"] Dec 03 19:25:15 crc kubenswrapper[4758]: I1203 19:25:15.128471 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rqqj"] Dec 03 19:25:16 crc kubenswrapper[4758]: I1203 19:25:16.274920 4758 generic.go:334] "Generic (PLEG): container finished" podID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerID="934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece" exitCode=0 Dec 03 19:25:16 crc kubenswrapper[4758]: I1203 19:25:16.275770 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerDied","Data":"934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece"} Dec 03 19:25:16 crc kubenswrapper[4758]: I1203 19:25:16.280859 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerStarted","Data":"46beb52331ddff8d931b8687cd991db4542d4230691044c0fa9a6fa2d888dcea"} Dec 03 19:25:16 crc kubenswrapper[4758]: I1203 19:25:16.283149 4758 generic.go:334] "Generic (PLEG): container finished" podID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerID="ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc" exitCode=0 Dec 03 19:25:16 crc kubenswrapper[4758]: I1203 19:25:16.283204 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xs5p7" event={"ID":"e0328a71-b1d5-412e-bd4c-776ff76c1157","Type":"ContainerDied","Data":"ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc"} Dec 03 19:25:16 crc kubenswrapper[4758]: I1203 19:25:16.283236 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xs5p7" event={"ID":"e0328a71-b1d5-412e-bd4c-776ff76c1157","Type":"ContainerStarted","Data":"f704cee4e767500d57a6d0bcf57dfcad809c6112b7a24120d16fbc3aa3052589"} Dec 03 19:25:18 crc kubenswrapper[4758]: I1203 19:25:18.307499 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerStarted","Data":"348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024"} Dec 03 19:25:18 crc kubenswrapper[4758]: I1203 19:25:18.312345 4758 generic.go:334] "Generic (PLEG): container finished" podID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerID="d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1" exitCode=0 Dec 03 19:25:18 crc kubenswrapper[4758]: I1203 19:25:18.312401 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xs5p7" event={"ID":"e0328a71-b1d5-412e-bd4c-776ff76c1157","Type":"ContainerDied","Data":"d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1"} Dec 03 19:25:19 crc kubenswrapper[4758]: E1203 19:25:19.979915 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22a1b4d0_e0b7_4894_bdc8_5c322b943e29.slice/crio-348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024.scope\": RecentStats: unable to find data in memory cache]" Dec 03 19:25:22 crc kubenswrapper[4758]: I1203 19:25:22.357955 4758 generic.go:334] "Generic (PLEG): container finished" podID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerID="348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024" exitCode=0 Dec 03 19:25:22 crc kubenswrapper[4758]: I1203 19:25:22.358039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerDied","Data":"348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024"} Dec 03 19:25:24 crc kubenswrapper[4758]: I1203 19:25:24.377934 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xs5p7" event={"ID":"e0328a71-b1d5-412e-bd4c-776ff76c1157","Type":"ContainerStarted","Data":"f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8"} Dec 03 19:25:24 crc kubenswrapper[4758]: I1203 19:25:24.380076 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerStarted","Data":"c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f"} Dec 03 19:25:24 crc kubenswrapper[4758]: I1203 19:25:24.408968 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xs5p7" podStartSLOduration=4.66531421 podStartE2EDuration="11.408945488s" podCreationTimestamp="2025-12-03 19:25:13 +0000 UTC" firstStartedPulling="2025-12-03 19:25:16.285633471 +0000 UTC m=+8971.487010342" lastFinishedPulling="2025-12-03 19:25:23.029264759 +0000 UTC m=+8978.230641620" observedRunningTime="2025-12-03 19:25:24.398825047 +0000 UTC m=+8979.600201928" watchObservedRunningTime="2025-12-03 19:25:24.408945488 +0000 UTC m=+8979.610322349" Dec 03 19:25:24 crc kubenswrapper[4758]: I1203 19:25:24.427153 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2rqqj" podStartSLOduration=3.606785012 podStartE2EDuration="10.427131176s" podCreationTimestamp="2025-12-03 19:25:14 +0000 UTC" firstStartedPulling="2025-12-03 19:25:16.278563672 +0000 UTC m=+8971.479940533" lastFinishedPulling="2025-12-03 19:25:23.098909836 +0000 UTC m=+8978.300286697" observedRunningTime="2025-12-03 19:25:24.421528235 +0000 UTC m=+8979.622905096" watchObservedRunningTime="2025-12-03 19:25:24.427131176 +0000 UTC m=+8979.628508037" Dec 03 19:25:24 crc kubenswrapper[4758]: I1203 19:25:24.540840 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:24 crc kubenswrapper[4758]: I1203 19:25:24.540978 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:25 crc kubenswrapper[4758]: I1203 19:25:25.585614 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2rqqj" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="registry-server" probeResult="failure" output=< Dec 03 19:25:25 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 19:25:25 crc kubenswrapper[4758]: > Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.365763 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.366435 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.418359 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.557821 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.599549 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.653138 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:34 crc kubenswrapper[4758]: I1203 19:25:34.663645 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xs5p7"] Dec 03 19:25:36 crc kubenswrapper[4758]: I1203 19:25:36.541579 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xs5p7" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="registry-server" containerID="cri-o://f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8" gracePeriod=2 Dec 03 19:25:36 crc kubenswrapper[4758]: I1203 19:25:36.861361 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2rqqj"] Dec 03 19:25:36 crc kubenswrapper[4758]: I1203 19:25:36.861931 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2rqqj" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="registry-server" containerID="cri-o://c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f" gracePeriod=2 Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.098784 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.231327 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-utilities\") pod \"e0328a71-b1d5-412e-bd4c-776ff76c1157\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.231544 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdsrb\" (UniqueName: \"kubernetes.io/projected/e0328a71-b1d5-412e-bd4c-776ff76c1157-kube-api-access-kdsrb\") pod \"e0328a71-b1d5-412e-bd4c-776ff76c1157\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.231743 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-catalog-content\") pod \"e0328a71-b1d5-412e-bd4c-776ff76c1157\" (UID: \"e0328a71-b1d5-412e-bd4c-776ff76c1157\") " Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.232083 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-utilities" (OuterVolumeSpecName: "utilities") pod "e0328a71-b1d5-412e-bd4c-776ff76c1157" (UID: "e0328a71-b1d5-412e-bd4c-776ff76c1157"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.238949 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0328a71-b1d5-412e-bd4c-776ff76c1157-kube-api-access-kdsrb" (OuterVolumeSpecName: "kube-api-access-kdsrb") pod "e0328a71-b1d5-412e-bd4c-776ff76c1157" (UID: "e0328a71-b1d5-412e-bd4c-776ff76c1157"). InnerVolumeSpecName "kube-api-access-kdsrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.253526 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0328a71-b1d5-412e-bd4c-776ff76c1157" (UID: "e0328a71-b1d5-412e-bd4c-776ff76c1157"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.334915 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.334952 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0328a71-b1d5-412e-bd4c-776ff76c1157-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.334966 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdsrb\" (UniqueName: \"kubernetes.io/projected/e0328a71-b1d5-412e-bd4c-776ff76c1157-kube-api-access-kdsrb\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.337934 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.436660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hw77\" (UniqueName: \"kubernetes.io/projected/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-kube-api-access-9hw77\") pod \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.437268 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-catalog-content\") pod \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.437304 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-utilities\") pod \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\" (UID: \"22a1b4d0-e0b7-4894-bdc8-5c322b943e29\") " Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.438027 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-utilities" (OuterVolumeSpecName: "utilities") pod "22a1b4d0-e0b7-4894-bdc8-5c322b943e29" (UID: "22a1b4d0-e0b7-4894-bdc8-5c322b943e29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.440151 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-kube-api-access-9hw77" (OuterVolumeSpecName: "kube-api-access-9hw77") pod "22a1b4d0-e0b7-4894-bdc8-5c322b943e29" (UID: "22a1b4d0-e0b7-4894-bdc8-5c322b943e29"). InnerVolumeSpecName "kube-api-access-9hw77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.538278 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22a1b4d0-e0b7-4894-bdc8-5c322b943e29" (UID: "22a1b4d0-e0b7-4894-bdc8-5c322b943e29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.540069 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hw77\" (UniqueName: \"kubernetes.io/projected/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-kube-api-access-9hw77\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.540097 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.540106 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22a1b4d0-e0b7-4894-bdc8-5c322b943e29-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.556815 4758 generic.go:334] "Generic (PLEG): container finished" podID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerID="c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f" exitCode=0 Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.556904 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerDied","Data":"c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f"} Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.556908 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rqqj" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.556966 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rqqj" event={"ID":"22a1b4d0-e0b7-4894-bdc8-5c322b943e29","Type":"ContainerDied","Data":"46beb52331ddff8d931b8687cd991db4542d4230691044c0fa9a6fa2d888dcea"} Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.556990 4758 scope.go:117] "RemoveContainer" containerID="c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.561091 4758 generic.go:334] "Generic (PLEG): container finished" podID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerID="f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8" exitCode=0 Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.561127 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xs5p7" event={"ID":"e0328a71-b1d5-412e-bd4c-776ff76c1157","Type":"ContainerDied","Data":"f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8"} Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.561150 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xs5p7" event={"ID":"e0328a71-b1d5-412e-bd4c-776ff76c1157","Type":"ContainerDied","Data":"f704cee4e767500d57a6d0bcf57dfcad809c6112b7a24120d16fbc3aa3052589"} Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.561184 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xs5p7" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.584109 4758 scope.go:117] "RemoveContainer" containerID="348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.610884 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2rqqj"] Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.640275 4758 scope.go:117] "RemoveContainer" containerID="934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.644882 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2rqqj"] Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.683522 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xs5p7"] Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.692063 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xs5p7"] Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.732750 4758 scope.go:117] "RemoveContainer" containerID="c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f" Dec 03 19:25:37 crc kubenswrapper[4758]: E1203 19:25:37.733300 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f\": container with ID starting with c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f not found: ID does not exist" containerID="c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.733372 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f"} err="failed to get container status \"c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f\": rpc error: code = NotFound desc = could not find container \"c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f\": container with ID starting with c28576983f948fed4bb5898ef5d9f69d1116f6520dd6e4d862ce25a7d525941f not found: ID does not exist" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.733412 4758 scope.go:117] "RemoveContainer" containerID="348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024" Dec 03 19:25:37 crc kubenswrapper[4758]: E1203 19:25:37.734083 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024\": container with ID starting with 348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024 not found: ID does not exist" containerID="348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.734132 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024"} err="failed to get container status \"348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024\": rpc error: code = NotFound desc = could not find container \"348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024\": container with ID starting with 348e528c4cd14b3c25c66194fc5c89d1fe2c0abfd1e29395f24d58ead6ce7024 not found: ID does not exist" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.734156 4758 scope.go:117] "RemoveContainer" containerID="934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece" Dec 03 19:25:37 crc kubenswrapper[4758]: E1203 19:25:37.734467 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece\": container with ID starting with 934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece not found: ID does not exist" containerID="934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.734486 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece"} err="failed to get container status \"934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece\": rpc error: code = NotFound desc = could not find container \"934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece\": container with ID starting with 934ed40dae554262110ba0ac2007a3e58fede97222ab5f3f80da84b4b6111ece not found: ID does not exist" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.734499 4758 scope.go:117] "RemoveContainer" containerID="f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.779965 4758 scope.go:117] "RemoveContainer" containerID="d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.804883 4758 scope.go:117] "RemoveContainer" containerID="ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.864562 4758 scope.go:117] "RemoveContainer" containerID="f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8" Dec 03 19:25:37 crc kubenswrapper[4758]: E1203 19:25:37.865084 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8\": container with ID starting with f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8 not found: ID does not exist" containerID="f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.865134 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8"} err="failed to get container status \"f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8\": rpc error: code = NotFound desc = could not find container \"f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8\": container with ID starting with f5b41cdd8c888ac0172384e85a06079e67bd48f34c8a5ecf0dd0dc67b556c1c8 not found: ID does not exist" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.865160 4758 scope.go:117] "RemoveContainer" containerID="d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1" Dec 03 19:25:37 crc kubenswrapper[4758]: E1203 19:25:37.865453 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1\": container with ID starting with d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1 not found: ID does not exist" containerID="d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.865484 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1"} err="failed to get container status \"d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1\": rpc error: code = NotFound desc = could not find container \"d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1\": container with ID starting with d781f9fda4b929311e6250a43921959ab77bfb37f9e969085eaa7b436278b9f1 not found: ID does not exist" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.865504 4758 scope.go:117] "RemoveContainer" containerID="ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc" Dec 03 19:25:37 crc kubenswrapper[4758]: E1203 19:25:37.865960 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc\": container with ID starting with ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc not found: ID does not exist" containerID="ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc" Dec 03 19:25:37 crc kubenswrapper[4758]: I1203 19:25:37.865984 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc"} err="failed to get container status \"ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc\": rpc error: code = NotFound desc = could not find container \"ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc\": container with ID starting with ceb08cbccd15fa49082f156d2a16d9ade7bb5a07b850df1c54a4cd5dd10629dc not found: ID does not exist" Dec 03 19:25:39 crc kubenswrapper[4758]: I1203 19:25:39.127219 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" path="/var/lib/kubelet/pods/22a1b4d0-e0b7-4894-bdc8-5c322b943e29/volumes" Dec 03 19:25:39 crc kubenswrapper[4758]: I1203 19:25:39.128350 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" path="/var/lib/kubelet/pods/e0328a71-b1d5-412e-bd4c-776ff76c1157/volumes" Dec 03 19:27:00 crc kubenswrapper[4758]: I1203 19:27:00.400510 4758 generic.go:334] "Generic (PLEG): container finished" podID="80659ad8-2251-46bb-853d-18ac040f6d84" containerID="a6a76c46b9e657ed9e07ad1f5086d383825228ec035fee46acbefcd1b564d694" exitCode=0 Dec 03 19:27:00 crc kubenswrapper[4758]: I1203 19:27:00.400575 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" event={"ID":"80659ad8-2251-46bb-853d-18ac040f6d84","Type":"ContainerDied","Data":"a6a76c46b9e657ed9e07ad1f5086d383825228ec035fee46acbefcd1b564d694"} Dec 03 19:27:01 crc kubenswrapper[4758]: I1203 19:27:01.898871 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.058771 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-agent-neutron-config-0\") pod \"80659ad8-2251-46bb-853d-18ac040f6d84\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.058819 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ssh-key\") pod \"80659ad8-2251-46bb-853d-18ac040f6d84\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.058930 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-955t8\" (UniqueName: \"kubernetes.io/projected/80659ad8-2251-46bb-853d-18ac040f6d84-kube-api-access-955t8\") pod \"80659ad8-2251-46bb-853d-18ac040f6d84\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.059136 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ceph\") pod \"80659ad8-2251-46bb-853d-18ac040f6d84\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.059206 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-combined-ca-bundle\") pod \"80659ad8-2251-46bb-853d-18ac040f6d84\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.059292 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-inventory\") pod \"80659ad8-2251-46bb-853d-18ac040f6d84\" (UID: \"80659ad8-2251-46bb-853d-18ac040f6d84\") " Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.065061 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ceph" (OuterVolumeSpecName: "ceph") pod "80659ad8-2251-46bb-853d-18ac040f6d84" (UID: "80659ad8-2251-46bb-853d-18ac040f6d84"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.069941 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "80659ad8-2251-46bb-853d-18ac040f6d84" (UID: "80659ad8-2251-46bb-853d-18ac040f6d84"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.070064 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80659ad8-2251-46bb-853d-18ac040f6d84-kube-api-access-955t8" (OuterVolumeSpecName: "kube-api-access-955t8") pod "80659ad8-2251-46bb-853d-18ac040f6d84" (UID: "80659ad8-2251-46bb-853d-18ac040f6d84"). InnerVolumeSpecName "kube-api-access-955t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.090944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "80659ad8-2251-46bb-853d-18ac040f6d84" (UID: "80659ad8-2251-46bb-853d-18ac040f6d84"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.095475 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80659ad8-2251-46bb-853d-18ac040f6d84" (UID: "80659ad8-2251-46bb-853d-18ac040f6d84"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.102294 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-inventory" (OuterVolumeSpecName: "inventory") pod "80659ad8-2251-46bb-853d-18ac040f6d84" (UID: "80659ad8-2251-46bb-853d-18ac040f6d84"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.162239 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.162272 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.162283 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.162296 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.162306 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80659ad8-2251-46bb-853d-18ac040f6d84-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.162315 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-955t8\" (UniqueName: \"kubernetes.io/projected/80659ad8-2251-46bb-853d-18ac040f6d84-kube-api-access-955t8\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.420764 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" event={"ID":"80659ad8-2251-46bb-853d-18ac040f6d84","Type":"ContainerDied","Data":"3e4997adfe3b28b149baf79bcedf9ee13b29448380e3c12ab9ad11e70c388220"} Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.421089 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e4997adfe3b28b149baf79bcedf9ee13b29448380e3c12ab9ad11e70c388220" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.420836 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-tzxnl" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.513044 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx"] Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515043 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="registry-server" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515089 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="registry-server" Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515110 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="registry-server" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515120 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="registry-server" Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515145 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="extract-content" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515154 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="extract-content" Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515180 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="extract-content" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515190 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="extract-content" Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515207 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="extract-utilities" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515215 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="extract-utilities" Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515233 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="extract-utilities" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515244 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="extract-utilities" Dec 03 19:27:02 crc kubenswrapper[4758]: E1203 19:27:02.515269 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80659ad8-2251-46bb-853d-18ac040f6d84" containerName="neutron-sriov-openstack-openstack-cell1" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515279 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="80659ad8-2251-46bb-853d-18ac040f6d84" containerName="neutron-sriov-openstack-openstack-cell1" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515700 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="80659ad8-2251-46bb-853d-18ac040f6d84" containerName="neutron-sriov-openstack-openstack-cell1" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515734 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0328a71-b1d5-412e-bd4c-776ff76c1157" containerName="registry-server" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.515764 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a1b4d0-e0b7-4894-bdc8-5c322b943e29" containerName="registry-server" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.516812 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.519393 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.519559 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.519703 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.520137 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.528774 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.547092 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx"] Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.570809 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm7cn\" (UniqueName: \"kubernetes.io/projected/4689d0b4-d192-4b64-bcb7-937098f1f2ee-kube-api-access-cm7cn\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.570883 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.570937 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.571079 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.571119 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.571142 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.676703 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm7cn\" (UniqueName: \"kubernetes.io/projected/4689d0b4-d192-4b64-bcb7-937098f1f2ee-kube-api-access-cm7cn\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.676813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.676850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.676944 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.676976 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.676997 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.681181 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.681294 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.682391 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.683592 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.684499 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.700271 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm7cn\" (UniqueName: \"kubernetes.io/projected/4689d0b4-d192-4b64-bcb7-937098f1f2ee-kube-api-access-cm7cn\") pod \"neutron-dhcp-openstack-openstack-cell1-hpqxx\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:02 crc kubenswrapper[4758]: I1203 19:27:02.847002 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:27:03 crc kubenswrapper[4758]: I1203 19:27:03.396224 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx"] Dec 03 19:27:03 crc kubenswrapper[4758]: I1203 19:27:03.431268 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" event={"ID":"4689d0b4-d192-4b64-bcb7-937098f1f2ee","Type":"ContainerStarted","Data":"1a6eff4508b74f052e83464c0dd2999142823eeadf8ad4639958d904e962dc7f"} Dec 03 19:27:04 crc kubenswrapper[4758]: I1203 19:27:04.444530 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" event={"ID":"4689d0b4-d192-4b64-bcb7-937098f1f2ee","Type":"ContainerStarted","Data":"5a69caaa2148fc5926c8234abbeac12dfcfe3c60adf51be9a6f7592982c8d3cc"} Dec 03 19:27:04 crc kubenswrapper[4758]: I1203 19:27:04.469048 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" podStartSLOduration=2.2832660799999998 podStartE2EDuration="2.469020302s" podCreationTimestamp="2025-12-03 19:27:02 +0000 UTC" firstStartedPulling="2025-12-03 19:27:03.393827363 +0000 UTC m=+9078.595204224" lastFinishedPulling="2025-12-03 19:27:03.579581585 +0000 UTC m=+9078.780958446" observedRunningTime="2025-12-03 19:27:04.459016513 +0000 UTC m=+9079.660393374" watchObservedRunningTime="2025-12-03 19:27:04.469020302 +0000 UTC m=+9079.670397163" Dec 03 19:27:11 crc kubenswrapper[4758]: I1203 19:27:11.395134 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:27:11 crc kubenswrapper[4758]: I1203 19:27:11.395737 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:27:41 crc kubenswrapper[4758]: I1203 19:27:41.394799 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:27:41 crc kubenswrapper[4758]: I1203 19:27:41.395325 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:28:11 crc kubenswrapper[4758]: I1203 19:28:11.394228 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:28:11 crc kubenswrapper[4758]: I1203 19:28:11.394804 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:28:11 crc kubenswrapper[4758]: I1203 19:28:11.394852 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:28:11 crc kubenswrapper[4758]: I1203 19:28:11.395718 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ee4ad7fd5bf09c0c2f45187fe086f2bafb84a762e68492bc44a65b75f10a229"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:28:11 crc kubenswrapper[4758]: I1203 19:28:11.395763 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://0ee4ad7fd5bf09c0c2f45187fe086f2bafb84a762e68492bc44a65b75f10a229" gracePeriod=600 Dec 03 19:28:12 crc kubenswrapper[4758]: I1203 19:28:12.149540 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="0ee4ad7fd5bf09c0c2f45187fe086f2bafb84a762e68492bc44a65b75f10a229" exitCode=0 Dec 03 19:28:12 crc kubenswrapper[4758]: I1203 19:28:12.149617 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"0ee4ad7fd5bf09c0c2f45187fe086f2bafb84a762e68492bc44a65b75f10a229"} Dec 03 19:28:12 crc kubenswrapper[4758]: I1203 19:28:12.150137 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce"} Dec 03 19:28:12 crc kubenswrapper[4758]: I1203 19:28:12.150162 4758 scope.go:117] "RemoveContainer" containerID="7cd05172be5208931f7cff01ac77adf844737b861911bb22dfbacef66b9dacc4" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.148055 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd"] Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.150864 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.162895 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd"] Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.163155 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.182340 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.195758 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgzxf\" (UniqueName: \"kubernetes.io/projected/b2b3069a-ae69-435b-aca9-c75b070c4900-kube-api-access-sgzxf\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.195867 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b3069a-ae69-435b-aca9-c75b070c4900-secret-volume\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.195893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b3069a-ae69-435b-aca9-c75b070c4900-config-volume\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.297546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgzxf\" (UniqueName: \"kubernetes.io/projected/b2b3069a-ae69-435b-aca9-c75b070c4900-kube-api-access-sgzxf\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.297701 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b3069a-ae69-435b-aca9-c75b070c4900-secret-volume\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.297739 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b3069a-ae69-435b-aca9-c75b070c4900-config-volume\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.298767 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b3069a-ae69-435b-aca9-c75b070c4900-config-volume\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.306812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b3069a-ae69-435b-aca9-c75b070c4900-secret-volume\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.314232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgzxf\" (UniqueName: \"kubernetes.io/projected/b2b3069a-ae69-435b-aca9-c75b070c4900-kube-api-access-sgzxf\") pod \"collect-profiles-29413170-bvqrd\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.473086 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:00 crc kubenswrapper[4758]: I1203 19:30:00.962268 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd"] Dec 03 19:30:01 crc kubenswrapper[4758]: I1203 19:30:01.248735 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" event={"ID":"b2b3069a-ae69-435b-aca9-c75b070c4900","Type":"ContainerStarted","Data":"07674c105a811f327fc7722c83105f1710e29d0d171195e5a494f0c61780edcc"} Dec 03 19:30:01 crc kubenswrapper[4758]: I1203 19:30:01.249118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" event={"ID":"b2b3069a-ae69-435b-aca9-c75b070c4900","Type":"ContainerStarted","Data":"8133e2076aeb100e372b27c37e8873007d850406ad227aa40e65c9155c110d19"} Dec 03 19:30:01 crc kubenswrapper[4758]: I1203 19:30:01.267659 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" podStartSLOduration=1.267636834 podStartE2EDuration="1.267636834s" podCreationTimestamp="2025-12-03 19:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:30:01.264347905 +0000 UTC m=+9256.465724776" watchObservedRunningTime="2025-12-03 19:30:01.267636834 +0000 UTC m=+9256.469013695" Dec 03 19:30:02 crc kubenswrapper[4758]: I1203 19:30:02.259306 4758 generic.go:334] "Generic (PLEG): container finished" podID="b2b3069a-ae69-435b-aca9-c75b070c4900" containerID="07674c105a811f327fc7722c83105f1710e29d0d171195e5a494f0c61780edcc" exitCode=0 Dec 03 19:30:02 crc kubenswrapper[4758]: I1203 19:30:02.259363 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" event={"ID":"b2b3069a-ae69-435b-aca9-c75b070c4900","Type":"ContainerDied","Data":"07674c105a811f327fc7722c83105f1710e29d0d171195e5a494f0c61780edcc"} Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.295419 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" event={"ID":"b2b3069a-ae69-435b-aca9-c75b070c4900","Type":"ContainerDied","Data":"8133e2076aeb100e372b27c37e8873007d850406ad227aa40e65c9155c110d19"} Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.295834 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8133e2076aeb100e372b27c37e8873007d850406ad227aa40e65c9155c110d19" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.445527 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.496641 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgzxf\" (UniqueName: \"kubernetes.io/projected/b2b3069a-ae69-435b-aca9-c75b070c4900-kube-api-access-sgzxf\") pod \"b2b3069a-ae69-435b-aca9-c75b070c4900\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.496759 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b3069a-ae69-435b-aca9-c75b070c4900-config-volume\") pod \"b2b3069a-ae69-435b-aca9-c75b070c4900\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.496873 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b3069a-ae69-435b-aca9-c75b070c4900-secret-volume\") pod \"b2b3069a-ae69-435b-aca9-c75b070c4900\" (UID: \"b2b3069a-ae69-435b-aca9-c75b070c4900\") " Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.499512 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2b3069a-ae69-435b-aca9-c75b070c4900-config-volume" (OuterVolumeSpecName: "config-volume") pod "b2b3069a-ae69-435b-aca9-c75b070c4900" (UID: "b2b3069a-ae69-435b-aca9-c75b070c4900"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.514049 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2b3069a-ae69-435b-aca9-c75b070c4900-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b2b3069a-ae69-435b-aca9-c75b070c4900" (UID: "b2b3069a-ae69-435b-aca9-c75b070c4900"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.519545 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b3069a-ae69-435b-aca9-c75b070c4900-kube-api-access-sgzxf" (OuterVolumeSpecName: "kube-api-access-sgzxf") pod "b2b3069a-ae69-435b-aca9-c75b070c4900" (UID: "b2b3069a-ae69-435b-aca9-c75b070c4900"). InnerVolumeSpecName "kube-api-access-sgzxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.602170 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgzxf\" (UniqueName: \"kubernetes.io/projected/b2b3069a-ae69-435b-aca9-c75b070c4900-kube-api-access-sgzxf\") on node \"crc\" DevicePath \"\"" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.602216 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b3069a-ae69-435b-aca9-c75b070c4900-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:30:04 crc kubenswrapper[4758]: I1203 19:30:04.602229 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b3069a-ae69-435b-aca9-c75b070c4900-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:30:05 crc kubenswrapper[4758]: I1203 19:30:05.307493 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-bvqrd" Dec 03 19:30:05 crc kubenswrapper[4758]: I1203 19:30:05.524526 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt"] Dec 03 19:30:05 crc kubenswrapper[4758]: I1203 19:30:05.534034 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-h4ztt"] Dec 03 19:30:07 crc kubenswrapper[4758]: I1203 19:30:07.128864 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a546bafd-cf70-4fb6-867d-e2dbf24b9064" path="/var/lib/kubelet/pods/a546bafd-cf70-4fb6-867d-e2dbf24b9064/volumes" Dec 03 19:30:11 crc kubenswrapper[4758]: I1203 19:30:11.396086 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:30:11 crc kubenswrapper[4758]: I1203 19:30:11.396531 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:30:18 crc kubenswrapper[4758]: I1203 19:30:18.386548 4758 scope.go:117] "RemoveContainer" containerID="fa63e34d478e1e941af28ff6ccb4398c0b34f316ccf37d25ba6153f4947a32c8" Dec 03 19:30:41 crc kubenswrapper[4758]: I1203 19:30:41.394301 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:30:41 crc kubenswrapper[4758]: I1203 19:30:41.394865 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.394654 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.395965 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.396591 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.399483 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.399599 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" gracePeriod=600 Dec 03 19:31:11 crc kubenswrapper[4758]: E1203 19:31:11.765291 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.983502 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" exitCode=0 Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.983588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce"} Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.983651 4758 scope.go:117] "RemoveContainer" containerID="0ee4ad7fd5bf09c0c2f45187fe086f2bafb84a762e68492bc44a65b75f10a229" Dec 03 19:31:11 crc kubenswrapper[4758]: I1203 19:31:11.986310 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:31:11 crc kubenswrapper[4758]: E1203 19:31:11.987100 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:31:27 crc kubenswrapper[4758]: I1203 19:31:27.115494 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:31:27 crc kubenswrapper[4758]: E1203 19:31:27.116739 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:31:39 crc kubenswrapper[4758]: I1203 19:31:39.114809 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:31:39 crc kubenswrapper[4758]: E1203 19:31:39.115723 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:31:52 crc kubenswrapper[4758]: I1203 19:31:52.115218 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:31:52 crc kubenswrapper[4758]: E1203 19:31:52.116118 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:32:05 crc kubenswrapper[4758]: I1203 19:32:05.124403 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:32:05 crc kubenswrapper[4758]: E1203 19:32:05.126527 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:32:16 crc kubenswrapper[4758]: I1203 19:32:16.114971 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:32:16 crc kubenswrapper[4758]: E1203 19:32:16.115716 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:32:30 crc kubenswrapper[4758]: I1203 19:32:30.738070 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="f81e7681-d858-4cf2-97dc-08a6e8c0b6f6" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 19:32:31 crc kubenswrapper[4758]: I1203 19:32:31.117134 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:32:31 crc kubenswrapper[4758]: E1203 19:32:31.117372 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:32:31 crc kubenswrapper[4758]: I1203 19:32:31.391094 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="f81e7681-d858-4cf2-97dc-08a6e8c0b6f6" containerName="ceilometer-central-agent" probeResult="failure" output=< Dec 03 19:32:31 crc kubenswrapper[4758]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 03 19:32:31 crc kubenswrapper[4758]: > Dec 03 19:32:34 crc kubenswrapper[4758]: I1203 19:32:34.824240 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="f81e7681-d858-4cf2-97dc-08a6e8c0b6f6" containerName="ceilometer-central-agent" probeResult="failure" output=< Dec 03 19:32:34 crc kubenswrapper[4758]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 03 19:32:34 crc kubenswrapper[4758]: > Dec 03 19:32:34 crc kubenswrapper[4758]: I1203 19:32:34.824854 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 03 19:32:34 crc kubenswrapper[4758]: I1203 19:32:34.825843 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"a643da5aad3d7b44feaa6b723d20522b0937a01b378d36bef4d3dc64f57b5896"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 03 19:32:34 crc kubenswrapper[4758]: I1203 19:32:34.825935 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f81e7681-d858-4cf2-97dc-08a6e8c0b6f6" containerName="ceilometer-central-agent" containerID="cri-o://a643da5aad3d7b44feaa6b723d20522b0937a01b378d36bef4d3dc64f57b5896" gracePeriod=30 Dec 03 19:32:36 crc kubenswrapper[4758]: I1203 19:32:36.323022 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:32:36 crc kubenswrapper[4758]: I1203 19:32:36.799000 4758 generic.go:334] "Generic (PLEG): container finished" podID="f81e7681-d858-4cf2-97dc-08a6e8c0b6f6" containerID="a643da5aad3d7b44feaa6b723d20522b0937a01b378d36bef4d3dc64f57b5896" exitCode=0 Dec 03 19:32:36 crc kubenswrapper[4758]: I1203 19:32:36.799048 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerDied","Data":"a643da5aad3d7b44feaa6b723d20522b0937a01b378d36bef4d3dc64f57b5896"} Dec 03 19:32:37 crc kubenswrapper[4758]: I1203 19:32:37.812161 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f81e7681-d858-4cf2-97dc-08a6e8c0b6f6","Type":"ContainerStarted","Data":"95f9795579f86a21367788cb707112f583fa9e050c3be2f4db4d960422300d27"} Dec 03 19:32:43 crc kubenswrapper[4758]: I1203 19:32:43.114429 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:32:43 crc kubenswrapper[4758]: E1203 19:32:43.115263 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:32:56 crc kubenswrapper[4758]: I1203 19:32:56.115035 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:32:56 crc kubenswrapper[4758]: E1203 19:32:56.115824 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:33:05 crc kubenswrapper[4758]: I1203 19:33:05.081388 4758 generic.go:334] "Generic (PLEG): container finished" podID="4689d0b4-d192-4b64-bcb7-937098f1f2ee" containerID="5a69caaa2148fc5926c8234abbeac12dfcfe3c60adf51be9a6f7592982c8d3cc" exitCode=0 Dec 03 19:33:05 crc kubenswrapper[4758]: I1203 19:33:05.081480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" event={"ID":"4689d0b4-d192-4b64-bcb7-937098f1f2ee","Type":"ContainerDied","Data":"5a69caaa2148fc5926c8234abbeac12dfcfe3c60adf51be9a6f7592982c8d3cc"} Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.590771 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.738546 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-agent-neutron-config-0\") pod \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.738802 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ssh-key\") pod \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.738869 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-inventory\") pod \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.738900 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ceph\") pod \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.739000 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm7cn\" (UniqueName: \"kubernetes.io/projected/4689d0b4-d192-4b64-bcb7-937098f1f2ee-kube-api-access-cm7cn\") pod \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.739035 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-combined-ca-bundle\") pod \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\" (UID: \"4689d0b4-d192-4b64-bcb7-937098f1f2ee\") " Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.744806 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ceph" (OuterVolumeSpecName: "ceph") pod "4689d0b4-d192-4b64-bcb7-937098f1f2ee" (UID: "4689d0b4-d192-4b64-bcb7-937098f1f2ee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.745104 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "4689d0b4-d192-4b64-bcb7-937098f1f2ee" (UID: "4689d0b4-d192-4b64-bcb7-937098f1f2ee"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.751139 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4689d0b4-d192-4b64-bcb7-937098f1f2ee-kube-api-access-cm7cn" (OuterVolumeSpecName: "kube-api-access-cm7cn") pod "4689d0b4-d192-4b64-bcb7-937098f1f2ee" (UID: "4689d0b4-d192-4b64-bcb7-937098f1f2ee"). InnerVolumeSpecName "kube-api-access-cm7cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.771588 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4689d0b4-d192-4b64-bcb7-937098f1f2ee" (UID: "4689d0b4-d192-4b64-bcb7-937098f1f2ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.774352 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "4689d0b4-d192-4b64-bcb7-937098f1f2ee" (UID: "4689d0b4-d192-4b64-bcb7-937098f1f2ee"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.776309 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-inventory" (OuterVolumeSpecName: "inventory") pod "4689d0b4-d192-4b64-bcb7-937098f1f2ee" (UID: "4689d0b4-d192-4b64-bcb7-937098f1f2ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.842392 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.842423 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.842452 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.842461 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.842471 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm7cn\" (UniqueName: \"kubernetes.io/projected/4689d0b4-d192-4b64-bcb7-937098f1f2ee-kube-api-access-cm7cn\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:06 crc kubenswrapper[4758]: I1203 19:33:06.842482 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4689d0b4-d192-4b64-bcb7-937098f1f2ee-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:07 crc kubenswrapper[4758]: I1203 19:33:07.103806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" event={"ID":"4689d0b4-d192-4b64-bcb7-937098f1f2ee","Type":"ContainerDied","Data":"1a6eff4508b74f052e83464c0dd2999142823eeadf8ad4639958d904e962dc7f"} Dec 03 19:33:07 crc kubenswrapper[4758]: I1203 19:33:07.103856 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a6eff4508b74f052e83464c0dd2999142823eeadf8ad4639958d904e962dc7f" Dec 03 19:33:07 crc kubenswrapper[4758]: I1203 19:33:07.103891 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hpqxx" Dec 03 19:33:11 crc kubenswrapper[4758]: I1203 19:33:11.114771 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:33:11 crc kubenswrapper[4758]: E1203 19:33:11.115566 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:33:19 crc kubenswrapper[4758]: I1203 19:33:19.860470 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 19:33:19 crc kubenswrapper[4758]: I1203 19:33:19.861224 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerName="nova-cell0-conductor-conductor" containerID="cri-o://a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" gracePeriod=30 Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.002381 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.002599 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="44772cdb-c8fc-49a8-a2b7-56a81cec54ce" containerName="nova-cell1-conductor-conductor" containerID="cri-o://9eeed526487510d141effa3293138fe5c24f0a9d0b697fefc042da231282688a" gracePeriod=30 Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.834753 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.835313 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerName="nova-scheduler-scheduler" containerID="cri-o://4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" gracePeriod=30 Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.853151 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.853385 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-log" containerID="cri-o://004d80a978ced39f198ca85260643f8bbfa0e18efa48068f072a2d9efe095e6f" gracePeriod=30 Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.853867 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-api" containerID="cri-o://707b6849d2d683a60141de7af8106e6805d937ae961eac28c6acd1631443d051" gracePeriod=30 Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.930003 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.930324 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-log" containerID="cri-o://7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741" gracePeriod=30 Dec 03 19:33:20 crc kubenswrapper[4758]: I1203 19:33:20.930506 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-metadata" containerID="cri-o://c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4" gracePeriod=30 Dec 03 19:33:21 crc kubenswrapper[4758]: I1203 19:33:21.259168 4758 generic.go:334] "Generic (PLEG): container finished" podID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerID="7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741" exitCode=143 Dec 03 19:33:21 crc kubenswrapper[4758]: I1203 19:33:21.259249 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ffb17f73-7b5f-4225-829b-53b13343ea77","Type":"ContainerDied","Data":"7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741"} Dec 03 19:33:21 crc kubenswrapper[4758]: I1203 19:33:21.262175 4758 generic.go:334] "Generic (PLEG): container finished" podID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerID="004d80a978ced39f198ca85260643f8bbfa0e18efa48068f072a2d9efe095e6f" exitCode=143 Dec 03 19:33:21 crc kubenswrapper[4758]: I1203 19:33:21.262242 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99ebca78-76cf-4cf6-817b-b9a421d8cfd3","Type":"ContainerDied","Data":"004d80a978ced39f198ca85260643f8bbfa0e18efa48068f072a2d9efe095e6f"} Dec 03 19:33:21 crc kubenswrapper[4758]: E1203 19:33:21.596192 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 19:33:21 crc kubenswrapper[4758]: E1203 19:33:21.598151 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 19:33:21 crc kubenswrapper[4758]: E1203 19:33:21.600900 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 19:33:21 crc kubenswrapper[4758]: E1203 19:33:21.601031 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerName="nova-scheduler-scheduler" Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.115320 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:33:22 crc kubenswrapper[4758]: E1203 19:33:22.115883 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.273354 4758 generic.go:334] "Generic (PLEG): container finished" podID="44772cdb-c8fc-49a8-a2b7-56a81cec54ce" containerID="9eeed526487510d141effa3293138fe5c24f0a9d0b697fefc042da231282688a" exitCode=0 Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.273405 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"44772cdb-c8fc-49a8-a2b7-56a81cec54ce","Type":"ContainerDied","Data":"9eeed526487510d141effa3293138fe5c24f0a9d0b697fefc042da231282688a"} Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.273439 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"44772cdb-c8fc-49a8-a2b7-56a81cec54ce","Type":"ContainerDied","Data":"4b09d17d5d3f96dbf428f076c5ea56a2c6cef5f6c641a87f441c70026658bf55"} Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.273453 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b09d17d5d3f96dbf428f076c5ea56a2c6cef5f6c641a87f441c70026658bf55" Dec 03 19:33:22 crc kubenswrapper[4758]: E1203 19:33:22.479413 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 19:33:22 crc kubenswrapper[4758]: E1203 19:33:22.481433 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 19:33:22 crc kubenswrapper[4758]: E1203 19:33:22.482585 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 19:33:22 crc kubenswrapper[4758]: E1203 19:33:22.482735 4758 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerName="nova-cell0-conductor-conductor" Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.880135 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.980381 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle\") pod \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.980611 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rbrv\" (UniqueName: \"kubernetes.io/projected/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-kube-api-access-2rbrv\") pod \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.980660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-config-data\") pod \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " Dec 03 19:33:22 crc kubenswrapper[4758]: I1203 19:33:22.988907 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-kube-api-access-2rbrv" (OuterVolumeSpecName: "kube-api-access-2rbrv") pod "44772cdb-c8fc-49a8-a2b7-56a81cec54ce" (UID: "44772cdb-c8fc-49a8-a2b7-56a81cec54ce"). InnerVolumeSpecName "kube-api-access-2rbrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:33:23 crc kubenswrapper[4758]: E1203 19:33:23.044609 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle podName:44772cdb-c8fc-49a8-a2b7-56a81cec54ce nodeName:}" failed. No retries permitted until 2025-12-03 19:33:23.544542465 +0000 UTC m=+9458.745919326 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle") pod "44772cdb-c8fc-49a8-a2b7-56a81cec54ce" (UID: "44772cdb-c8fc-49a8-a2b7-56a81cec54ce") : error deleting /var/lib/kubelet/pods/44772cdb-c8fc-49a8-a2b7-56a81cec54ce/volume-subpaths: remove /var/lib/kubelet/pods/44772cdb-c8fc-49a8-a2b7-56a81cec54ce/volume-subpaths: no such file or directory Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.048701 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-config-data" (OuterVolumeSpecName: "config-data") pod "44772cdb-c8fc-49a8-a2b7-56a81cec54ce" (UID: "44772cdb-c8fc-49a8-a2b7-56a81cec54ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.082918 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rbrv\" (UniqueName: \"kubernetes.io/projected/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-kube-api-access-2rbrv\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.082951 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.283019 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.592887 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle\") pod \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\" (UID: \"44772cdb-c8fc-49a8-a2b7-56a81cec54ce\") " Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.603903 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44772cdb-c8fc-49a8-a2b7-56a81cec54ce" (UID: "44772cdb-c8fc-49a8-a2b7-56a81cec54ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.696505 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44772cdb-c8fc-49a8-a2b7-56a81cec54ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.940655 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.958877 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.970954 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 19:33:23 crc kubenswrapper[4758]: E1203 19:33:23.971554 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44772cdb-c8fc-49a8-a2b7-56a81cec54ce" containerName="nova-cell1-conductor-conductor" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.971583 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="44772cdb-c8fc-49a8-a2b7-56a81cec54ce" containerName="nova-cell1-conductor-conductor" Dec 03 19:33:23 crc kubenswrapper[4758]: E1203 19:33:23.971603 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4689d0b4-d192-4b64-bcb7-937098f1f2ee" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.971611 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4689d0b4-d192-4b64-bcb7-937098f1f2ee" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 03 19:33:23 crc kubenswrapper[4758]: E1203 19:33:23.971655 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b3069a-ae69-435b-aca9-c75b070c4900" containerName="collect-profiles" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.971664 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b3069a-ae69-435b-aca9-c75b070c4900" containerName="collect-profiles" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.971945 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4689d0b4-d192-4b64-bcb7-937098f1f2ee" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.971983 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="44772cdb-c8fc-49a8-a2b7-56a81cec54ce" containerName="nova-cell1-conductor-conductor" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.972011 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b3069a-ae69-435b-aca9-c75b070c4900" containerName="collect-profiles" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.972784 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.976909 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 19:33:23 crc kubenswrapper[4758]: I1203 19:33:23.983598 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.013435 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc8eb0-3634-4e45-9d75-5023116c4585-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.013529 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxpdn\" (UniqueName: \"kubernetes.io/projected/e8bc8eb0-3634-4e45-9d75-5023116c4585-kube-api-access-mxpdn\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.013742 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc8eb0-3634-4e45-9d75-5023116c4585-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.081849 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": read tcp 10.217.0.2:50678->10.217.1.85:8775: read: connection reset by peer" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.081941 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": read tcp 10.217.0.2:50684->10.217.1.85:8775: read: connection reset by peer" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.115159 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc8eb0-3634-4e45-9d75-5023116c4585-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.115252 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc8eb0-3634-4e45-9d75-5023116c4585-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.115297 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxpdn\" (UniqueName: \"kubernetes.io/projected/e8bc8eb0-3634-4e45-9d75-5023116c4585-kube-api-access-mxpdn\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.118972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc8eb0-3634-4e45-9d75-5023116c4585-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.126516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc8eb0-3634-4e45-9d75-5023116c4585-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.130317 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxpdn\" (UniqueName: \"kubernetes.io/projected/e8bc8eb0-3634-4e45-9d75-5023116c4585-kube-api-access-mxpdn\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc8eb0-3634-4e45-9d75-5023116c4585\") " pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.292245 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:24 crc kubenswrapper[4758]: I1203 19:33:24.844129 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.107304 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.127545 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44772cdb-c8fc-49a8-a2b7-56a81cec54ce" path="/var/lib/kubelet/pods/44772cdb-c8fc-49a8-a2b7-56a81cec54ce/volumes" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.146288 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb17f73-7b5f-4225-829b-53b13343ea77-logs\") pod \"ffb17f73-7b5f-4225-829b-53b13343ea77\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.146342 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-combined-ca-bundle\") pod \"ffb17f73-7b5f-4225-829b-53b13343ea77\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.146577 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjwb4\" (UniqueName: \"kubernetes.io/projected/ffb17f73-7b5f-4225-829b-53b13343ea77-kube-api-access-kjwb4\") pod \"ffb17f73-7b5f-4225-829b-53b13343ea77\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.146655 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-config-data\") pod \"ffb17f73-7b5f-4225-829b-53b13343ea77\" (UID: \"ffb17f73-7b5f-4225-829b-53b13343ea77\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.151628 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffb17f73-7b5f-4225-829b-53b13343ea77-kube-api-access-kjwb4" (OuterVolumeSpecName: "kube-api-access-kjwb4") pod "ffb17f73-7b5f-4225-829b-53b13343ea77" (UID: "ffb17f73-7b5f-4225-829b-53b13343ea77"). InnerVolumeSpecName "kube-api-access-kjwb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.249246 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjwb4\" (UniqueName: \"kubernetes.io/projected/ffb17f73-7b5f-4225-829b-53b13343ea77-kube-api-access-kjwb4\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.264120 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffb17f73-7b5f-4225-829b-53b13343ea77-logs" (OuterVolumeSpecName: "logs") pod "ffb17f73-7b5f-4225-829b-53b13343ea77" (UID: "ffb17f73-7b5f-4225-829b-53b13343ea77"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.270069 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-config-data" (OuterVolumeSpecName: "config-data") pod "ffb17f73-7b5f-4225-829b-53b13343ea77" (UID: "ffb17f73-7b5f-4225-829b-53b13343ea77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.306382 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffb17f73-7b5f-4225-829b-53b13343ea77" (UID: "ffb17f73-7b5f-4225-829b-53b13343ea77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.311018 4758 generic.go:334] "Generic (PLEG): container finished" podID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerID="c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4" exitCode=0 Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.311193 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.319874 4758 generic.go:334] "Generic (PLEG): container finished" podID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerID="707b6849d2d683a60141de7af8106e6805d937ae961eac28c6acd1631443d051" exitCode=0 Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.351520 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.351558 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb17f73-7b5f-4225-829b-53b13343ea77-logs\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.351568 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb17f73-7b5f-4225-829b-53b13343ea77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.389672 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ffb17f73-7b5f-4225-829b-53b13343ea77","Type":"ContainerDied","Data":"c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4"} Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.389834 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ffb17f73-7b5f-4225-829b-53b13343ea77","Type":"ContainerDied","Data":"4d38561873a64e6dc5b78e857a1bfbbc9f7a1e1a320d6cad644d5e5a597acf5d"} Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.389867 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99ebca78-76cf-4cf6-817b-b9a421d8cfd3","Type":"ContainerDied","Data":"707b6849d2d683a60141de7af8106e6805d937ae961eac28c6acd1631443d051"} Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.389912 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8bc8eb0-3634-4e45-9d75-5023116c4585","Type":"ContainerStarted","Data":"47350785d8554f2bac860dc608e17e780cd1b9090cbbb3349b6f843ecb5cb7d3"} Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.389940 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8bc8eb0-3634-4e45-9d75-5023116c4585","Type":"ContainerStarted","Data":"c73120457b6225a0eed2c19db9be89aa9d2f9eaa716d4640ccd64392d3c8db28"} Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.389974 4758 scope.go:117] "RemoveContainer" containerID="c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.430602 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.442338 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.464311 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 19:33:25 crc kubenswrapper[4758]: E1203 19:33:25.464955 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-metadata" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.464975 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-metadata" Dec 03 19:33:25 crc kubenswrapper[4758]: E1203 19:33:25.464989 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-log" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.464997 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-log" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.465295 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-log" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.465334 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" containerName="nova-metadata-metadata" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.466804 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.469694 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.488763 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.499986 4758 scope.go:117] "RemoveContainer" containerID="7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.535097 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.563098 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a675e1-3969-4c03-8026-5c7206ed8dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.563516 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a675e1-3969-4c03-8026-5c7206ed8dee-config-data\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.563676 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7smz\" (UniqueName: \"kubernetes.io/projected/01a675e1-3969-4c03-8026-5c7206ed8dee-kube-api-access-t7smz\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.563989 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01a675e1-3969-4c03-8026-5c7206ed8dee-logs\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.572669 4758 scope.go:117] "RemoveContainer" containerID="c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4" Dec 03 19:33:25 crc kubenswrapper[4758]: E1203 19:33:25.577602 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4\": container with ID starting with c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4 not found: ID does not exist" containerID="c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.577646 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4"} err="failed to get container status \"c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4\": rpc error: code = NotFound desc = could not find container \"c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4\": container with ID starting with c28a7dc6ed6379902e05d1274c8ac4a3602f78f9e56b07f688f02ead91f545b4 not found: ID does not exist" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.577672 4758 scope.go:117] "RemoveContainer" containerID="7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741" Dec 03 19:33:25 crc kubenswrapper[4758]: E1203 19:33:25.578123 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741\": container with ID starting with 7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741 not found: ID does not exist" containerID="7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.578168 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741"} err="failed to get container status \"7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741\": rpc error: code = NotFound desc = could not find container \"7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741\": container with ID starting with 7ff26fcd2897860ddbfaeb46ab199c93381f3e21a80fc9bd671819104a406741 not found: ID does not exist" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.665883 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-config-data\") pod \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666013 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl6vr\" (UniqueName: \"kubernetes.io/projected/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-kube-api-access-bl6vr\") pod \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666081 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-combined-ca-bundle\") pod \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666305 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-logs\") pod \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\" (UID: \"99ebca78-76cf-4cf6-817b-b9a421d8cfd3\") " Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666634 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a675e1-3969-4c03-8026-5c7206ed8dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666755 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a675e1-3969-4c03-8026-5c7206ed8dee-config-data\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666789 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7smz\" (UniqueName: \"kubernetes.io/projected/01a675e1-3969-4c03-8026-5c7206ed8dee-kube-api-access-t7smz\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.666885 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01a675e1-3969-4c03-8026-5c7206ed8dee-logs\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.667268 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01a675e1-3969-4c03-8026-5c7206ed8dee-logs\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.671530 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-kube-api-access-bl6vr" (OuterVolumeSpecName: "kube-api-access-bl6vr") pod "99ebca78-76cf-4cf6-817b-b9a421d8cfd3" (UID: "99ebca78-76cf-4cf6-817b-b9a421d8cfd3"). InnerVolumeSpecName "kube-api-access-bl6vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.673156 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-logs" (OuterVolumeSpecName: "logs") pod "99ebca78-76cf-4cf6-817b-b9a421d8cfd3" (UID: "99ebca78-76cf-4cf6-817b-b9a421d8cfd3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.676095 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a675e1-3969-4c03-8026-5c7206ed8dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.676194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a675e1-3969-4c03-8026-5c7206ed8dee-config-data\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.695576 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7smz\" (UniqueName: \"kubernetes.io/projected/01a675e1-3969-4c03-8026-5c7206ed8dee-kube-api-access-t7smz\") pod \"nova-metadata-0\" (UID: \"01a675e1-3969-4c03-8026-5c7206ed8dee\") " pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.705916 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99ebca78-76cf-4cf6-817b-b9a421d8cfd3" (UID: "99ebca78-76cf-4cf6-817b-b9a421d8cfd3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.729796 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-config-data" (OuterVolumeSpecName: "config-data") pod "99ebca78-76cf-4cf6-817b-b9a421d8cfd3" (UID: "99ebca78-76cf-4cf6-817b-b9a421d8cfd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.752073 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v"] Dec 03 19:33:25 crc kubenswrapper[4758]: E1203 19:33:25.752503 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-log" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.752522 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-log" Dec 03 19:33:25 crc kubenswrapper[4758]: E1203 19:33:25.752547 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-api" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.752555 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-api" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.752807 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-api" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.752835 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" containerName="nova-api-log" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.753702 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.758751 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.758840 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.758847 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.759010 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-trmp7" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.759033 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.759162 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.759282 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.767927 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v"] Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.768455 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-logs\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.768528 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.768595 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl6vr\" (UniqueName: \"kubernetes.io/projected/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-kube-api-access-bl6vr\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.768650 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ebca78-76cf-4cf6-817b-b9a421d8cfd3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.806845 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.870844 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871223 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871262 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871336 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871371 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871433 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.871544 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vh8d\" (UniqueName: \"kubernetes.io/projected/f5a21acf-8141-463b-855f-959e09815eee-kube-api-access-6vh8d\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.872120 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.872190 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973519 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973577 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973605 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973673 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973724 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973758 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973793 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973841 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.973929 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vh8d\" (UniqueName: \"kubernetes.io/projected/f5a21acf-8141-463b-855f-959e09815eee-kube-api-access-6vh8d\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.977395 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.977402 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.982618 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.986119 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.986274 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.986936 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.988572 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.990092 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.992495 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:25 crc kubenswrapper[4758]: I1203 19:33:25.992806 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.001247 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vh8d\" (UniqueName: \"kubernetes.io/projected/f5a21acf-8141-463b-855f-959e09815eee-kube-api-access-6vh8d\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.081335 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.322167 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.349151 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.349231 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99ebca78-76cf-4cf6-817b-b9a421d8cfd3","Type":"ContainerDied","Data":"1a5432a170aa183623673c3fd33edd1efb2c079a36a1763d035a1fdeafae0bc5"} Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.349285 4758 scope.go:117] "RemoveContainer" containerID="707b6849d2d683a60141de7af8106e6805d937ae961eac28c6acd1631443d051" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.349393 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.372245 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.372225162 podStartE2EDuration="3.372225162s" podCreationTimestamp="2025-12-03 19:33:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:33:26.369018926 +0000 UTC m=+9461.570395787" watchObservedRunningTime="2025-12-03 19:33:26.372225162 +0000 UTC m=+9461.573602023" Dec 03 19:33:26 crc kubenswrapper[4758]: W1203 19:33:26.545226 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5a21acf_8141_463b_855f_959e09815eee.slice/crio-94483d8c4c5170000ab313ff133b37bb37430e550e47dba64267fe5b36486750 WatchSource:0}: Error finding container 94483d8c4c5170000ab313ff133b37bb37430e550e47dba64267fe5b36486750: Status 404 returned error can't find the container with id 94483d8c4c5170000ab313ff133b37bb37430e550e47dba64267fe5b36486750 Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.546305 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v"] Dec 03 19:33:26 crc kubenswrapper[4758]: E1203 19:33:26.594151 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb is running failed: container process not found" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 19:33:26 crc kubenswrapper[4758]: E1203 19:33:26.594555 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb is running failed: container process not found" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 19:33:26 crc kubenswrapper[4758]: E1203 19:33:26.594838 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb is running failed: container process not found" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 19:33:26 crc kubenswrapper[4758]: E1203 19:33:26.594887 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerName="nova-scheduler-scheduler" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.627307 4758 scope.go:117] "RemoveContainer" containerID="004d80a978ced39f198ca85260643f8bbfa0e18efa48068f072a2d9efe095e6f" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.650536 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.665879 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.707670 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.712804 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.725066 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.748924 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.799663 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e024a3-38fa-4d28-9276-b4cf52206784-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.799733 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e024a3-38fa-4d28-9276-b4cf52206784-config-data\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.799826 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxtqh\" (UniqueName: \"kubernetes.io/projected/c5e024a3-38fa-4d28-9276-b4cf52206784-kube-api-access-vxtqh\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.799848 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5e024a3-38fa-4d28-9276-b4cf52206784-logs\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.904477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e024a3-38fa-4d28-9276-b4cf52206784-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.904556 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e024a3-38fa-4d28-9276-b4cf52206784-config-data\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.904659 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxtqh\" (UniqueName: \"kubernetes.io/projected/c5e024a3-38fa-4d28-9276-b4cf52206784-kube-api-access-vxtqh\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.904711 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5e024a3-38fa-4d28-9276-b4cf52206784-logs\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.905575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5e024a3-38fa-4d28-9276-b4cf52206784-logs\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.924941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e024a3-38fa-4d28-9276-b4cf52206784-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.925238 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e024a3-38fa-4d28-9276-b4cf52206784-config-data\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:26 crc kubenswrapper[4758]: I1203 19:33:26.926037 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxtqh\" (UniqueName: \"kubernetes.io/projected/c5e024a3-38fa-4d28-9276-b4cf52206784-kube-api-access-vxtqh\") pod \"nova-api-0\" (UID: \"c5e024a3-38fa-4d28-9276-b4cf52206784\") " pod="openstack/nova-api-0" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.058673 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.147200 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99ebca78-76cf-4cf6-817b-b9a421d8cfd3" path="/var/lib/kubelet/pods/99ebca78-76cf-4cf6-817b-b9a421d8cfd3/volumes" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.150133 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffb17f73-7b5f-4225-829b-53b13343ea77" path="/var/lib/kubelet/pods/ffb17f73-7b5f-4225-829b-53b13343ea77/volumes" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.389224 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" event={"ID":"f5a21acf-8141-463b-855f-959e09815eee","Type":"ContainerStarted","Data":"0327cc624979e13ff8080f1c08ef63fbbdda659d5a44f6dfbb9808e5a456f5f0"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.389571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" event={"ID":"f5a21acf-8141-463b-855f-959e09815eee","Type":"ContainerStarted","Data":"94483d8c4c5170000ab313ff133b37bb37430e550e47dba64267fe5b36486750"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.390334 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.413112 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"01a675e1-3969-4c03-8026-5c7206ed8dee","Type":"ContainerStarted","Data":"072b37a9baf6ba83e3c1af05abf2a2f67d7c773ae66ac23994b5ad821b69492b"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.413167 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"01a675e1-3969-4c03-8026-5c7206ed8dee","Type":"ContainerStarted","Data":"28288e4b7aaa63aee428adec7081b8c4f0f838b99f10188da629caadc741b4f2"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.413181 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"01a675e1-3969-4c03-8026-5c7206ed8dee","Type":"ContainerStarted","Data":"d0ece6243407e2a1de6fd8a6b8d4a78e41ec827e611c93f19aa6438e3f23a838"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.419580 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" exitCode=0 Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.419660 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d1f5f4e9-0563-4d1a-b686-51d7657dfc76","Type":"ContainerDied","Data":"a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.421260 4758 generic.go:334] "Generic (PLEG): container finished" podID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" exitCode=0 Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.421320 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58aed963-c66b-445d-8c0f-28dc1c0694f8","Type":"ContainerDied","Data":"4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.421345 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58aed963-c66b-445d-8c0f-28dc1c0694f8","Type":"ContainerDied","Data":"713d10a5476970a53ad24afd1d0a1b8dbbe2d8f8d37511c8a709dc7cab34f816"} Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.421366 4758 scope.go:117] "RemoveContainer" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.421504 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.465104 4758 scope.go:117] "RemoveContainer" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.466673 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb\": container with ID starting with 4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb not found: ID does not exist" containerID="4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.466744 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb"} err="failed to get container status \"4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb\": rpc error: code = NotFound desc = could not find container \"4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb\": container with ID starting with 4f882266d49e21ab7845962058af692cd6b9b0b2faed5f04b847f67386e0f5eb not found: ID does not exist" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.468171 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" podStartSLOduration=2.207636526 podStartE2EDuration="2.46814936s" podCreationTimestamp="2025-12-03 19:33:25 +0000 UTC" firstStartedPulling="2025-12-03 19:33:26.547797829 +0000 UTC m=+9461.749174690" lastFinishedPulling="2025-12-03 19:33:26.808310663 +0000 UTC m=+9462.009687524" observedRunningTime="2025-12-03 19:33:27.427579267 +0000 UTC m=+9462.628956138" watchObservedRunningTime="2025-12-03 19:33:27.46814936 +0000 UTC m=+9462.669526221" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.480793 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.480771509 podStartE2EDuration="2.480771509s" podCreationTimestamp="2025-12-03 19:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:33:27.44885599 +0000 UTC m=+9462.650232851" watchObservedRunningTime="2025-12-03 19:33:27.480771509 +0000 UTC m=+9462.682148370" Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.481836 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d is running failed: container process not found" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.485780 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d is running failed: container process not found" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.488894 4758 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d is running failed: container process not found" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.488974 4758 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerName="nova-cell0-conductor-conductor" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.535809 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-combined-ca-bundle\") pod \"58aed963-c66b-445d-8c0f-28dc1c0694f8\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.536117 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-config-data\") pod \"58aed963-c66b-445d-8c0f-28dc1c0694f8\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.536163 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4zkq\" (UniqueName: \"kubernetes.io/projected/58aed963-c66b-445d-8c0f-28dc1c0694f8-kube-api-access-h4zkq\") pod \"58aed963-c66b-445d-8c0f-28dc1c0694f8\" (UID: \"58aed963-c66b-445d-8c0f-28dc1c0694f8\") " Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.565877 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58aed963-c66b-445d-8c0f-28dc1c0694f8-kube-api-access-h4zkq" (OuterVolumeSpecName: "kube-api-access-h4zkq") pod "58aed963-c66b-445d-8c0f-28dc1c0694f8" (UID: "58aed963-c66b-445d-8c0f-28dc1c0694f8"). InnerVolumeSpecName "kube-api-access-h4zkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.591856 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-config-data" (OuterVolumeSpecName: "config-data") pod "58aed963-c66b-445d-8c0f-28dc1c0694f8" (UID: "58aed963-c66b-445d-8c0f-28dc1c0694f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.641470 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.641507 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4zkq\" (UniqueName: \"kubernetes.io/projected/58aed963-c66b-445d-8c0f-28dc1c0694f8-kube-api-access-h4zkq\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.670950 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58aed963-c66b-445d-8c0f-28dc1c0694f8" (UID: "58aed963-c66b-445d-8c0f-28dc1c0694f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:27 crc kubenswrapper[4758]: W1203 19:33:27.726532 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5e024a3_38fa_4d28_9276_b4cf52206784.slice/crio-e135eaad4b4b554727658b9581d309b33738dd6d687d0bacf90cae6f20421597 WatchSource:0}: Error finding container e135eaad4b4b554727658b9581d309b33738dd6d687d0bacf90cae6f20421597: Status 404 returned error can't find the container with id e135eaad4b4b554727658b9581d309b33738dd6d687d0bacf90cae6f20421597 Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.744024 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58aed963-c66b-445d-8c0f-28dc1c0694f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.745521 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.764724 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.777325 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.825117 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.846710 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-config-data\") pod \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.846803 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxgth\" (UniqueName: \"kubernetes.io/projected/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-kube-api-access-wxgth\") pod \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.847177 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-combined-ca-bundle\") pod \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\" (UID: \"d1f5f4e9-0563-4d1a-b686-51d7657dfc76\") " Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.860554 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-kube-api-access-wxgth" (OuterVolumeSpecName: "kube-api-access-wxgth") pod "d1f5f4e9-0563-4d1a-b686-51d7657dfc76" (UID: "d1f5f4e9-0563-4d1a-b686-51d7657dfc76"). InnerVolumeSpecName "kube-api-access-wxgth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.911938 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.914595 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerName="nova-cell0-conductor-conductor" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.914639 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerName="nova-cell0-conductor-conductor" Dec 03 19:33:27 crc kubenswrapper[4758]: E1203 19:33:27.914656 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerName="nova-scheduler-scheduler" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.914664 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerName="nova-scheduler-scheduler" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.915099 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" containerName="nova-scheduler-scheduler" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.915174 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" containerName="nova-cell0-conductor-conductor" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.917642 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.925397 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1f5f4e9-0563-4d1a-b686-51d7657dfc76" (UID: "d1f5f4e9-0563-4d1a-b686-51d7657dfc76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.926932 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-config-data" (OuterVolumeSpecName: "config-data") pod "d1f5f4e9-0563-4d1a-b686-51d7657dfc76" (UID: "d1f5f4e9-0563-4d1a-b686-51d7657dfc76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.927230 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.953331 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.953378 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:27 crc kubenswrapper[4758]: I1203 19:33:27.953393 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxgth\" (UniqueName: \"kubernetes.io/projected/d1f5f4e9-0563-4d1a-b686-51d7657dfc76-kube-api-access-wxgth\") on node \"crc\" DevicePath \"\"" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.007974 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.055347 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3860513-f195-43f5-af71-26cb0e5a37a8-config-data\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.055809 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3860513-f195-43f5-af71-26cb0e5a37a8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.055900 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqf4h\" (UniqueName: \"kubernetes.io/projected/d3860513-f195-43f5-af71-26cb0e5a37a8-kube-api-access-pqf4h\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.158009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3860513-f195-43f5-af71-26cb0e5a37a8-config-data\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.158134 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3860513-f195-43f5-af71-26cb0e5a37a8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.158213 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqf4h\" (UniqueName: \"kubernetes.io/projected/d3860513-f195-43f5-af71-26cb0e5a37a8-kube-api-access-pqf4h\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.163187 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3860513-f195-43f5-af71-26cb0e5a37a8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.163662 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3860513-f195-43f5-af71-26cb0e5a37a8-config-data\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.189446 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqf4h\" (UniqueName: \"kubernetes.io/projected/d3860513-f195-43f5-af71-26cb0e5a37a8-kube-api-access-pqf4h\") pod \"nova-scheduler-0\" (UID: \"d3860513-f195-43f5-af71-26cb0e5a37a8\") " pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.281111 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.474962 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5e024a3-38fa-4d28-9276-b4cf52206784","Type":"ContainerStarted","Data":"4c67d01453e2237569862d4ccbc332f4bcb51056d8c90329f9b7296f5256554f"} Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.475454 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5e024a3-38fa-4d28-9276-b4cf52206784","Type":"ContainerStarted","Data":"e135eaad4b4b554727658b9581d309b33738dd6d687d0bacf90cae6f20421597"} Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.478137 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d1f5f4e9-0563-4d1a-b686-51d7657dfc76","Type":"ContainerDied","Data":"29f981a03cac529ac6ea332443ce93f97696a5bc939c5b454718edeb961436e2"} Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.478190 4758 scope.go:117] "RemoveContainer" containerID="a940908c09377f9423b777f663b4ba62c0a2ed8cba5485ec7c172fde170d708d" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.479404 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.580019 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.623732 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.636413 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.638040 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.640664 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.649109 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.778815 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56369ab2-3a2d-4164-8801-de6ad870b6fa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.779420 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56369ab2-3a2d-4164-8801-de6ad870b6fa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.779475 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxzrj\" (UniqueName: \"kubernetes.io/projected/56369ab2-3a2d-4164-8801-de6ad870b6fa-kube-api-access-vxzrj\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.864790 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 19:33:28 crc kubenswrapper[4758]: W1203 19:33:28.868015 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3860513_f195_43f5_af71_26cb0e5a37a8.slice/crio-49a54d2e5e75c60745f55721b1992f9a8c043ec5bd6136637c77e8ad94ef9e6a WatchSource:0}: Error finding container 49a54d2e5e75c60745f55721b1992f9a8c043ec5bd6136637c77e8ad94ef9e6a: Status 404 returned error can't find the container with id 49a54d2e5e75c60745f55721b1992f9a8c043ec5bd6136637c77e8ad94ef9e6a Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.881061 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56369ab2-3a2d-4164-8801-de6ad870b6fa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.881130 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxzrj\" (UniqueName: \"kubernetes.io/projected/56369ab2-3a2d-4164-8801-de6ad870b6fa-kube-api-access-vxzrj\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.881187 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56369ab2-3a2d-4164-8801-de6ad870b6fa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.885050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56369ab2-3a2d-4164-8801-de6ad870b6fa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.885324 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56369ab2-3a2d-4164-8801-de6ad870b6fa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.902268 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxzrj\" (UniqueName: \"kubernetes.io/projected/56369ab2-3a2d-4164-8801-de6ad870b6fa-kube-api-access-vxzrj\") pod \"nova-cell0-conductor-0\" (UID: \"56369ab2-3a2d-4164-8801-de6ad870b6fa\") " pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:28 crc kubenswrapper[4758]: I1203 19:33:28.980297 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.144391 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58aed963-c66b-445d-8c0f-28dc1c0694f8" path="/var/lib/kubelet/pods/58aed963-c66b-445d-8c0f-28dc1c0694f8/volumes" Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.145482 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f5f4e9-0563-4d1a-b686-51d7657dfc76" path="/var/lib/kubelet/pods/d1f5f4e9-0563-4d1a-b686-51d7657dfc76/volumes" Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.506644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5e024a3-38fa-4d28-9276-b4cf52206784","Type":"ContainerStarted","Data":"161b8b99fb12c9e409952f0b006ef3df49b0eeae763439413604c3952bbc3494"} Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.511954 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3860513-f195-43f5-af71-26cb0e5a37a8","Type":"ContainerStarted","Data":"d36447dca87202cf99a87c6ae6b3987cf0d0c870f04c5002e455862b6ba77bbd"} Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.512017 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3860513-f195-43f5-af71-26cb0e5a37a8","Type":"ContainerStarted","Data":"49a54d2e5e75c60745f55721b1992f9a8c043ec5bd6136637c77e8ad94ef9e6a"} Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.546093 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.559641 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.559618501 podStartE2EDuration="3.559618501s" podCreationTimestamp="2025-12-03 19:33:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:33:29.534358311 +0000 UTC m=+9464.735735172" watchObservedRunningTime="2025-12-03 19:33:29.559618501 +0000 UTC m=+9464.760995372" Dec 03 19:33:29 crc kubenswrapper[4758]: I1203 19:33:29.572232 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5722132 podStartE2EDuration="2.5722132s" podCreationTimestamp="2025-12-03 19:33:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:33:29.554765651 +0000 UTC m=+9464.756142512" watchObservedRunningTime="2025-12-03 19:33:29.5722132 +0000 UTC m=+9464.773590061" Dec 03 19:33:30 crc kubenswrapper[4758]: I1203 19:33:30.525513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"56369ab2-3a2d-4164-8801-de6ad870b6fa","Type":"ContainerStarted","Data":"324218c83fab7cf084f28589282cce5752928e58d0949e9611e063fc8cffe831"} Dec 03 19:33:30 crc kubenswrapper[4758]: I1203 19:33:30.525924 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:30 crc kubenswrapper[4758]: I1203 19:33:30.525940 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"56369ab2-3a2d-4164-8801-de6ad870b6fa","Type":"ContainerStarted","Data":"a1ab04fdf4d9f7cd81b99950093c966a6171af6ed293d552c7bd53a17aa4b274"} Dec 03 19:33:30 crc kubenswrapper[4758]: I1203 19:33:30.547492 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.547475189 podStartE2EDuration="2.547475189s" podCreationTimestamp="2025-12-03 19:33:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:33:30.541399995 +0000 UTC m=+9465.742776856" watchObservedRunningTime="2025-12-03 19:33:30.547475189 +0000 UTC m=+9465.748852040" Dec 03 19:33:30 crc kubenswrapper[4758]: I1203 19:33:30.807233 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 19:33:30 crc kubenswrapper[4758]: I1203 19:33:30.807370 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 19:33:33 crc kubenswrapper[4758]: I1203 19:33:33.282011 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 19:33:34 crc kubenswrapper[4758]: I1203 19:33:34.114525 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:33:34 crc kubenswrapper[4758]: E1203 19:33:34.114827 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:33:34 crc kubenswrapper[4758]: I1203 19:33:34.332030 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 19:33:35 crc kubenswrapper[4758]: I1203 19:33:35.807730 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 19:33:35 crc kubenswrapper[4758]: I1203 19:33:35.808099 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 19:33:36 crc kubenswrapper[4758]: I1203 19:33:36.890949 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="01a675e1-3969-4c03-8026-5c7206ed8dee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.194:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:33:36 crc kubenswrapper[4758]: I1203 19:33:36.891008 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="01a675e1-3969-4c03-8026-5c7206ed8dee" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.194:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:33:37 crc kubenswrapper[4758]: I1203 19:33:37.062740 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 19:33:37 crc kubenswrapper[4758]: I1203 19:33:37.067387 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 19:33:38 crc kubenswrapper[4758]: I1203 19:33:38.145941 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c5e024a3-38fa-4d28-9276-b4cf52206784" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:33:38 crc kubenswrapper[4758]: I1203 19:33:38.146503 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c5e024a3-38fa-4d28-9276-b4cf52206784" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:33:38 crc kubenswrapper[4758]: I1203 19:33:38.281267 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 19:33:38 crc kubenswrapper[4758]: I1203 19:33:38.312697 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 19:33:38 crc kubenswrapper[4758]: I1203 19:33:38.656016 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 19:33:39 crc kubenswrapper[4758]: I1203 19:33:39.022418 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 19:33:45 crc kubenswrapper[4758]: I1203 19:33:45.809521 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 19:33:45 crc kubenswrapper[4758]: I1203 19:33:45.810164 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 19:33:45 crc kubenswrapper[4758]: I1203 19:33:45.812984 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 19:33:45 crc kubenswrapper[4758]: I1203 19:33:45.813042 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 19:33:47 crc kubenswrapper[4758]: I1203 19:33:47.063673 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 19:33:47 crc kubenswrapper[4758]: I1203 19:33:47.064299 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 19:33:47 crc kubenswrapper[4758]: I1203 19:33:47.064776 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 19:33:47 crc kubenswrapper[4758]: I1203 19:33:47.064808 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 19:33:47 crc kubenswrapper[4758]: I1203 19:33:47.067449 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 19:33:47 crc kubenswrapper[4758]: I1203 19:33:47.077264 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 19:33:48 crc kubenswrapper[4758]: I1203 19:33:48.114456 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:33:48 crc kubenswrapper[4758]: E1203 19:33:48.115172 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:34:02 crc kubenswrapper[4758]: I1203 19:34:02.115091 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:34:02 crc kubenswrapper[4758]: E1203 19:34:02.116985 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:34:17 crc kubenswrapper[4758]: I1203 19:34:17.114568 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:34:17 crc kubenswrapper[4758]: E1203 19:34:17.115377 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:34:18 crc kubenswrapper[4758]: I1203 19:34:18.501092 4758 scope.go:117] "RemoveContainer" containerID="9eeed526487510d141effa3293138fe5c24f0a9d0b697fefc042da231282688a" Dec 03 19:34:32 crc kubenswrapper[4758]: I1203 19:34:32.114242 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:34:32 crc kubenswrapper[4758]: E1203 19:34:32.115296 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:34:44 crc kubenswrapper[4758]: I1203 19:34:44.115808 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:34:44 crc kubenswrapper[4758]: E1203 19:34:44.117113 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.571300 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xr6jh"] Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.615568 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.627320 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6spqc\" (UniqueName: \"kubernetes.io/projected/fc062fee-cdab-4317-8d18-060ffbcd983f-kube-api-access-6spqc\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.627671 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-catalog-content\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.627941 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-utilities\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.634121 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xr6jh"] Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.729958 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6spqc\" (UniqueName: \"kubernetes.io/projected/fc062fee-cdab-4317-8d18-060ffbcd983f-kube-api-access-6spqc\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.730009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-catalog-content\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.730069 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-utilities\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.730541 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-catalog-content\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.731476 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-utilities\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.750178 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6spqc\" (UniqueName: \"kubernetes.io/projected/fc062fee-cdab-4317-8d18-060ffbcd983f-kube-api-access-6spqc\") pod \"certified-operators-xr6jh\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:47 crc kubenswrapper[4758]: I1203 19:34:47.941828 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:48 crc kubenswrapper[4758]: I1203 19:34:48.489389 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xr6jh"] Dec 03 19:34:48 crc kubenswrapper[4758]: W1203 19:34:48.491132 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc062fee_cdab_4317_8d18_060ffbcd983f.slice/crio-b2257fb6cfe94531ed8dd7bd0d3dae894e497468fdc68273e1174495886608ea WatchSource:0}: Error finding container b2257fb6cfe94531ed8dd7bd0d3dae894e497468fdc68273e1174495886608ea: Status 404 returned error can't find the container with id b2257fb6cfe94531ed8dd7bd0d3dae894e497468fdc68273e1174495886608ea Dec 03 19:34:49 crc kubenswrapper[4758]: I1203 19:34:49.337949 4758 generic.go:334] "Generic (PLEG): container finished" podID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerID="b22764317f4e2908089e4ee776a4ca76be39d76865234bf44c82f81c5adf7464" exitCode=0 Dec 03 19:34:49 crc kubenswrapper[4758]: I1203 19:34:49.339546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerDied","Data":"b22764317f4e2908089e4ee776a4ca76be39d76865234bf44c82f81c5adf7464"} Dec 03 19:34:49 crc kubenswrapper[4758]: I1203 19:34:49.339621 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerStarted","Data":"b2257fb6cfe94531ed8dd7bd0d3dae894e497468fdc68273e1174495886608ea"} Dec 03 19:34:50 crc kubenswrapper[4758]: I1203 19:34:50.353766 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerStarted","Data":"60e4bbf36048bf623888201548c0e5934990eafab202fcfed8194eb175b4b891"} Dec 03 19:34:51 crc kubenswrapper[4758]: I1203 19:34:51.367202 4758 generic.go:334] "Generic (PLEG): container finished" podID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerID="60e4bbf36048bf623888201548c0e5934990eafab202fcfed8194eb175b4b891" exitCode=0 Dec 03 19:34:51 crc kubenswrapper[4758]: I1203 19:34:51.367275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerDied","Data":"60e4bbf36048bf623888201548c0e5934990eafab202fcfed8194eb175b4b891"} Dec 03 19:34:54 crc kubenswrapper[4758]: I1203 19:34:54.399784 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerStarted","Data":"a837e1c8507ac52f2469f2392917c0682c712b1d3f8b02984ad8838c89fe8163"} Dec 03 19:34:54 crc kubenswrapper[4758]: I1203 19:34:54.430034 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xr6jh" podStartSLOduration=4.939728102 podStartE2EDuration="7.429988561s" podCreationTimestamp="2025-12-03 19:34:47 +0000 UTC" firstStartedPulling="2025-12-03 19:34:49.346941043 +0000 UTC m=+9544.548317924" lastFinishedPulling="2025-12-03 19:34:51.837201512 +0000 UTC m=+9547.038578383" observedRunningTime="2025-12-03 19:34:54.421170524 +0000 UTC m=+9549.622547385" watchObservedRunningTime="2025-12-03 19:34:54.429988561 +0000 UTC m=+9549.631365422" Dec 03 19:34:57 crc kubenswrapper[4758]: I1203 19:34:57.943529 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:57 crc kubenswrapper[4758]: I1203 19:34:57.944157 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:57 crc kubenswrapper[4758]: I1203 19:34:57.995280 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:58 crc kubenswrapper[4758]: I1203 19:34:58.115607 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:34:58 crc kubenswrapper[4758]: E1203 19:34:58.116021 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:34:58 crc kubenswrapper[4758]: I1203 19:34:58.483279 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:34:58 crc kubenswrapper[4758]: I1203 19:34:58.531462 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xr6jh"] Dec 03 19:35:00 crc kubenswrapper[4758]: I1203 19:35:00.462937 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xr6jh" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="registry-server" containerID="cri-o://a837e1c8507ac52f2469f2392917c0682c712b1d3f8b02984ad8838c89fe8163" gracePeriod=2 Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.480241 4758 generic.go:334] "Generic (PLEG): container finished" podID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerID="a837e1c8507ac52f2469f2392917c0682c712b1d3f8b02984ad8838c89fe8163" exitCode=0 Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.480326 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerDied","Data":"a837e1c8507ac52f2469f2392917c0682c712b1d3f8b02984ad8838c89fe8163"} Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.480741 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xr6jh" event={"ID":"fc062fee-cdab-4317-8d18-060ffbcd983f","Type":"ContainerDied","Data":"b2257fb6cfe94531ed8dd7bd0d3dae894e497468fdc68273e1174495886608ea"} Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.480753 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2257fb6cfe94531ed8dd7bd0d3dae894e497468fdc68273e1174495886608ea" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.480523 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.647940 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-catalog-content\") pod \"fc062fee-cdab-4317-8d18-060ffbcd983f\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.648070 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6spqc\" (UniqueName: \"kubernetes.io/projected/fc062fee-cdab-4317-8d18-060ffbcd983f-kube-api-access-6spqc\") pod \"fc062fee-cdab-4317-8d18-060ffbcd983f\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.648145 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-utilities\") pod \"fc062fee-cdab-4317-8d18-060ffbcd983f\" (UID: \"fc062fee-cdab-4317-8d18-060ffbcd983f\") " Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.649450 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-utilities" (OuterVolumeSpecName: "utilities") pod "fc062fee-cdab-4317-8d18-060ffbcd983f" (UID: "fc062fee-cdab-4317-8d18-060ffbcd983f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.654878 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc062fee-cdab-4317-8d18-060ffbcd983f-kube-api-access-6spqc" (OuterVolumeSpecName: "kube-api-access-6spqc") pod "fc062fee-cdab-4317-8d18-060ffbcd983f" (UID: "fc062fee-cdab-4317-8d18-060ffbcd983f"). InnerVolumeSpecName "kube-api-access-6spqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.700787 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc062fee-cdab-4317-8d18-060ffbcd983f" (UID: "fc062fee-cdab-4317-8d18-060ffbcd983f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.751483 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.751744 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc062fee-cdab-4317-8d18-060ffbcd983f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:01 crc kubenswrapper[4758]: I1203 19:35:01.751863 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6spqc\" (UniqueName: \"kubernetes.io/projected/fc062fee-cdab-4317-8d18-060ffbcd983f-kube-api-access-6spqc\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:02 crc kubenswrapper[4758]: I1203 19:35:02.490501 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xr6jh" Dec 03 19:35:02 crc kubenswrapper[4758]: I1203 19:35:02.527721 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xr6jh"] Dec 03 19:35:02 crc kubenswrapper[4758]: I1203 19:35:02.538825 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xr6jh"] Dec 03 19:35:03 crc kubenswrapper[4758]: I1203 19:35:03.128747 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" path="/var/lib/kubelet/pods/fc062fee-cdab-4317-8d18-060ffbcd983f/volumes" Dec 03 19:35:13 crc kubenswrapper[4758]: I1203 19:35:13.114809 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:35:13 crc kubenswrapper[4758]: E1203 19:35:13.115825 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:35:28 crc kubenswrapper[4758]: I1203 19:35:28.115306 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:35:28 crc kubenswrapper[4758]: E1203 19:35:28.116099 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.975093 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7mjws"] Dec 03 19:35:40 crc kubenswrapper[4758]: E1203 19:35:40.976258 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="extract-content" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.976276 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="extract-content" Dec 03 19:35:40 crc kubenswrapper[4758]: E1203 19:35:40.976303 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="extract-utilities" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.976311 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="extract-utilities" Dec 03 19:35:40 crc kubenswrapper[4758]: E1203 19:35:40.976329 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="registry-server" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.976339 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="registry-server" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.976604 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc062fee-cdab-4317-8d18-060ffbcd983f" containerName="registry-server" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.978505 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:40 crc kubenswrapper[4758]: I1203 19:35:40.987359 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7mjws"] Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.115846 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:35:41 crc kubenswrapper[4758]: E1203 19:35:41.116344 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.132226 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-catalog-content\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.132412 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-utilities\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.132491 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd5jq\" (UniqueName: \"kubernetes.io/projected/afbcbb14-3465-47a7-9dc8-4952d41aa169-kube-api-access-qd5jq\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.235057 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd5jq\" (UniqueName: \"kubernetes.io/projected/afbcbb14-3465-47a7-9dc8-4952d41aa169-kube-api-access-qd5jq\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.235202 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-catalog-content\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.235975 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-catalog-content\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.236528 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-utilities\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.236983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-utilities\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.263511 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd5jq\" (UniqueName: \"kubernetes.io/projected/afbcbb14-3465-47a7-9dc8-4952d41aa169-kube-api-access-qd5jq\") pod \"community-operators-7mjws\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:41 crc kubenswrapper[4758]: I1203 19:35:41.304515 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:42 crc kubenswrapper[4758]: I1203 19:35:42.030303 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7mjws"] Dec 03 19:35:42 crc kubenswrapper[4758]: I1203 19:35:42.950847 4758 generic.go:334] "Generic (PLEG): container finished" podID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerID="bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a" exitCode=0 Dec 03 19:35:42 crc kubenswrapper[4758]: I1203 19:35:42.951028 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7mjws" event={"ID":"afbcbb14-3465-47a7-9dc8-4952d41aa169","Type":"ContainerDied","Data":"bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a"} Dec 03 19:35:42 crc kubenswrapper[4758]: I1203 19:35:42.951164 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7mjws" event={"ID":"afbcbb14-3465-47a7-9dc8-4952d41aa169","Type":"ContainerStarted","Data":"2b78e30afb518797e18d2414aeecc18c0c52455a278bddb98e4f7eb1722686cb"} Dec 03 19:35:47 crc kubenswrapper[4758]: I1203 19:35:47.010114 4758 generic.go:334] "Generic (PLEG): container finished" podID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerID="329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034" exitCode=0 Dec 03 19:35:47 crc kubenswrapper[4758]: I1203 19:35:47.010546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7mjws" event={"ID":"afbcbb14-3465-47a7-9dc8-4952d41aa169","Type":"ContainerDied","Data":"329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034"} Dec 03 19:35:48 crc kubenswrapper[4758]: I1203 19:35:48.024048 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7mjws" event={"ID":"afbcbb14-3465-47a7-9dc8-4952d41aa169","Type":"ContainerStarted","Data":"7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412"} Dec 03 19:35:48 crc kubenswrapper[4758]: I1203 19:35:48.048401 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7mjws" podStartSLOduration=3.559577049 podStartE2EDuration="8.048385394s" podCreationTimestamp="2025-12-03 19:35:40 +0000 UTC" firstStartedPulling="2025-12-03 19:35:42.954294388 +0000 UTC m=+9598.155671249" lastFinishedPulling="2025-12-03 19:35:47.443102733 +0000 UTC m=+9602.644479594" observedRunningTime="2025-12-03 19:35:48.042963029 +0000 UTC m=+9603.244339900" watchObservedRunningTime="2025-12-03 19:35:48.048385394 +0000 UTC m=+9603.249762255" Dec 03 19:35:51 crc kubenswrapper[4758]: I1203 19:35:51.305458 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:51 crc kubenswrapper[4758]: I1203 19:35:51.306037 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:52 crc kubenswrapper[4758]: I1203 19:35:52.174435 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:52 crc kubenswrapper[4758]: I1203 19:35:52.225504 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:52 crc kubenswrapper[4758]: I1203 19:35:52.412050 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7mjws"] Dec 03 19:35:53 crc kubenswrapper[4758]: I1203 19:35:53.114667 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:35:53 crc kubenswrapper[4758]: E1203 19:35:53.115091 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.084933 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7mjws" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="registry-server" containerID="cri-o://7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412" gracePeriod=2 Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.595264 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.766580 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-catalog-content\") pod \"afbcbb14-3465-47a7-9dc8-4952d41aa169\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.766674 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd5jq\" (UniqueName: \"kubernetes.io/projected/afbcbb14-3465-47a7-9dc8-4952d41aa169-kube-api-access-qd5jq\") pod \"afbcbb14-3465-47a7-9dc8-4952d41aa169\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.766727 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-utilities\") pod \"afbcbb14-3465-47a7-9dc8-4952d41aa169\" (UID: \"afbcbb14-3465-47a7-9dc8-4952d41aa169\") " Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.767624 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-utilities" (OuterVolumeSpecName: "utilities") pod "afbcbb14-3465-47a7-9dc8-4952d41aa169" (UID: "afbcbb14-3465-47a7-9dc8-4952d41aa169"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.780872 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afbcbb14-3465-47a7-9dc8-4952d41aa169-kube-api-access-qd5jq" (OuterVolumeSpecName: "kube-api-access-qd5jq") pod "afbcbb14-3465-47a7-9dc8-4952d41aa169" (UID: "afbcbb14-3465-47a7-9dc8-4952d41aa169"). InnerVolumeSpecName "kube-api-access-qd5jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.841977 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afbcbb14-3465-47a7-9dc8-4952d41aa169" (UID: "afbcbb14-3465-47a7-9dc8-4952d41aa169"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.876014 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.876063 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd5jq\" (UniqueName: \"kubernetes.io/projected/afbcbb14-3465-47a7-9dc8-4952d41aa169-kube-api-access-qd5jq\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:54 crc kubenswrapper[4758]: I1203 19:35:54.876079 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afbcbb14-3465-47a7-9dc8-4952d41aa169-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.099207 4758 generic.go:334] "Generic (PLEG): container finished" podID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerID="7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412" exitCode=0 Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.099252 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7mjws" event={"ID":"afbcbb14-3465-47a7-9dc8-4952d41aa169","Type":"ContainerDied","Data":"7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412"} Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.099280 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7mjws" event={"ID":"afbcbb14-3465-47a7-9dc8-4952d41aa169","Type":"ContainerDied","Data":"2b78e30afb518797e18d2414aeecc18c0c52455a278bddb98e4f7eb1722686cb"} Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.099283 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7mjws" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.099295 4758 scope.go:117] "RemoveContainer" containerID="7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.146930 4758 scope.go:117] "RemoveContainer" containerID="329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.156906 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7mjws"] Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.186546 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7mjws"] Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.204321 4758 scope.go:117] "RemoveContainer" containerID="bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.234522 4758 scope.go:117] "RemoveContainer" containerID="7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412" Dec 03 19:35:55 crc kubenswrapper[4758]: E1203 19:35:55.234919 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412\": container with ID starting with 7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412 not found: ID does not exist" containerID="7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.235007 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412"} err="failed to get container status \"7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412\": rpc error: code = NotFound desc = could not find container \"7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412\": container with ID starting with 7d21d67d4ba9da1642c8b56d5500b4c4ebf42b19f0604cb63b01326d2e7d4412 not found: ID does not exist" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.235081 4758 scope.go:117] "RemoveContainer" containerID="329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034" Dec 03 19:35:55 crc kubenswrapper[4758]: E1203 19:35:55.235571 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034\": container with ID starting with 329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034 not found: ID does not exist" containerID="329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.235611 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034"} err="failed to get container status \"329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034\": rpc error: code = NotFound desc = could not find container \"329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034\": container with ID starting with 329428b4e912d2d210ca6b92eab9dbe828b0877ec6d2878aa306507f4849a034 not found: ID does not exist" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.235636 4758 scope.go:117] "RemoveContainer" containerID="bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a" Dec 03 19:35:55 crc kubenswrapper[4758]: E1203 19:35:55.235868 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a\": container with ID starting with bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a not found: ID does not exist" containerID="bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a" Dec 03 19:35:55 crc kubenswrapper[4758]: I1203 19:35:55.240936 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a"} err="failed to get container status \"bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a\": rpc error: code = NotFound desc = could not find container \"bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a\": container with ID starting with bc89091975c34f76a7c9b0c0ea742c395cb73003e531e28db2e57060991ae03a not found: ID does not exist" Dec 03 19:35:57 crc kubenswrapper[4758]: I1203 19:35:57.127412 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" path="/var/lib/kubelet/pods/afbcbb14-3465-47a7-9dc8-4952d41aa169/volumes" Dec 03 19:36:06 crc kubenswrapper[4758]: I1203 19:36:06.114222 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:36:06 crc kubenswrapper[4758]: E1203 19:36:06.115159 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:36:18 crc kubenswrapper[4758]: I1203 19:36:18.114460 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:36:19 crc kubenswrapper[4758]: I1203 19:36:19.353590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"21a90d7bea8cfab67e8c91f5063700f03e750127eee664fa2caca4a804f6cffa"} Dec 03 19:36:29 crc kubenswrapper[4758]: I1203 19:36:29.446936 4758 generic.go:334] "Generic (PLEG): container finished" podID="f5a21acf-8141-463b-855f-959e09815eee" containerID="0327cc624979e13ff8080f1c08ef63fbbdda659d5a44f6dfbb9808e5a456f5f0" exitCode=0 Dec 03 19:36:29 crc kubenswrapper[4758]: I1203 19:36:29.447110 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" event={"ID":"f5a21acf-8141-463b-855f-959e09815eee","Type":"ContainerDied","Data":"0327cc624979e13ff8080f1c08ef63fbbdda659d5a44f6dfbb9808e5a456f5f0"} Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.009307 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.057900 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-0\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058363 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ssh-key\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058430 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-combined-ca-bundle\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058459 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ceph\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058568 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vh8d\" (UniqueName: \"kubernetes.io/projected/f5a21acf-8141-463b-855f-959e09815eee-kube-api-access-6vh8d\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058615 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-1\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058636 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-1\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058739 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-0\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058831 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-0\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058907 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-inventory\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.058984 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-1\") pod \"f5a21acf-8141-463b-855f-959e09815eee\" (UID: \"f5a21acf-8141-463b-855f-959e09815eee\") " Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.065403 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.066284 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a21acf-8141-463b-855f-959e09815eee-kube-api-access-6vh8d" (OuterVolumeSpecName: "kube-api-access-6vh8d") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "kube-api-access-6vh8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.082996 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ceph" (OuterVolumeSpecName: "ceph") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.092450 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.102353 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.101867 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.104721 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.105820 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.110989 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.113474 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-inventory" (OuterVolumeSpecName: "inventory") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.132254 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "f5a21acf-8141-463b-855f-959e09815eee" (UID: "f5a21acf-8141-463b-855f-959e09815eee"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161546 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161621 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161639 4758 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161652 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161667 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161695 4758 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161708 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/f5a21acf-8141-463b-855f-959e09815eee-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161718 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161730 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161743 4758 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5a21acf-8141-463b-855f-959e09815eee-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.161754 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vh8d\" (UniqueName: \"kubernetes.io/projected/f5a21acf-8141-463b-855f-959e09815eee-kube-api-access-6vh8d\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.469211 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" event={"ID":"f5a21acf-8141-463b-855f-959e09815eee","Type":"ContainerDied","Data":"94483d8c4c5170000ab313ff133b37bb37430e550e47dba64267fe5b36486750"} Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.469261 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94483d8c4c5170000ab313ff133b37bb37430e550e47dba64267fe5b36486750" Dec 03 19:36:31 crc kubenswrapper[4758]: I1203 19:36:31.469335 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.079044 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7664c"] Dec 03 19:36:35 crc kubenswrapper[4758]: E1203 19:36:35.080585 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="extract-content" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.080606 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="extract-content" Dec 03 19:36:35 crc kubenswrapper[4758]: E1203 19:36:35.080631 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="registry-server" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.080639 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="registry-server" Dec 03 19:36:35 crc kubenswrapper[4758]: E1203 19:36:35.080669 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="extract-utilities" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.080699 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="extract-utilities" Dec 03 19:36:35 crc kubenswrapper[4758]: E1203 19:36:35.080720 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a21acf-8141-463b-855f-959e09815eee" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.080730 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a21acf-8141-463b-855f-959e09815eee" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.081056 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="afbcbb14-3465-47a7-9dc8-4952d41aa169" containerName="registry-server" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.081076 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a21acf-8141-463b-855f-959e09815eee" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.083007 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.094754 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7664c"] Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.245278 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-utilities\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.245337 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-catalog-content\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.245705 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbhsx\" (UniqueName: \"kubernetes.io/projected/91673073-3678-461e-8ed3-24b827f2a252-kube-api-access-jbhsx\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.348146 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-utilities\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.348203 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-catalog-content\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.348294 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbhsx\" (UniqueName: \"kubernetes.io/projected/91673073-3678-461e-8ed3-24b827f2a252-kube-api-access-jbhsx\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.348747 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-utilities\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.348782 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-catalog-content\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.379202 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbhsx\" (UniqueName: \"kubernetes.io/projected/91673073-3678-461e-8ed3-24b827f2a252-kube-api-access-jbhsx\") pod \"redhat-operators-7664c\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.443064 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:35 crc kubenswrapper[4758]: I1203 19:36:35.951999 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7664c"] Dec 03 19:36:36 crc kubenswrapper[4758]: W1203 19:36:36.539593 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91673073_3678_461e_8ed3_24b827f2a252.slice/crio-270cde1f5d733fdc2e570c6e25c184891bf28e9f47a83b5df9238b6462d00705 WatchSource:0}: Error finding container 270cde1f5d733fdc2e570c6e25c184891bf28e9f47a83b5df9238b6462d00705: Status 404 returned error can't find the container with id 270cde1f5d733fdc2e570c6e25c184891bf28e9f47a83b5df9238b6462d00705 Dec 03 19:36:37 crc kubenswrapper[4758]: I1203 19:36:37.537523 4758 generic.go:334] "Generic (PLEG): container finished" podID="91673073-3678-461e-8ed3-24b827f2a252" containerID="22287e750fc817267c3d5867b06d5e74ac4708e494535b3364c9ae07754ac877" exitCode=0 Dec 03 19:36:37 crc kubenswrapper[4758]: I1203 19:36:37.537846 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerDied","Data":"22287e750fc817267c3d5867b06d5e74ac4708e494535b3364c9ae07754ac877"} Dec 03 19:36:37 crc kubenswrapper[4758]: I1203 19:36:37.537875 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerStarted","Data":"270cde1f5d733fdc2e570c6e25c184891bf28e9f47a83b5df9238b6462d00705"} Dec 03 19:36:38 crc kubenswrapper[4758]: I1203 19:36:38.547854 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerStarted","Data":"cda70aefffa7c97e2d75c7b628b8fadd2748247d6eb6d12f0e40e003ae526307"} Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.584350 4758 generic.go:334] "Generic (PLEG): container finished" podID="91673073-3678-461e-8ed3-24b827f2a252" containerID="cda70aefffa7c97e2d75c7b628b8fadd2748247d6eb6d12f0e40e003ae526307" exitCode=0 Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.584440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerDied","Data":"cda70aefffa7c97e2d75c7b628b8fadd2748247d6eb6d12f0e40e003ae526307"} Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.827993 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vtdbz"] Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.830407 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.855505 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtdbz"] Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.904835 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd7tr\" (UniqueName: \"kubernetes.io/projected/f05e6807-d0c3-467c-862a-ca31b573f8ab-kube-api-access-cd7tr\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.905235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-utilities\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:42 crc kubenswrapper[4758]: I1203 19:36:42.905297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-catalog-content\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.006917 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd7tr\" (UniqueName: \"kubernetes.io/projected/f05e6807-d0c3-467c-862a-ca31b573f8ab-kube-api-access-cd7tr\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.007015 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-utilities\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.007074 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-catalog-content\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.007588 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-utilities\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.007656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-catalog-content\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.026702 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd7tr\" (UniqueName: \"kubernetes.io/projected/f05e6807-d0c3-467c-862a-ca31b573f8ab-kube-api-access-cd7tr\") pod \"redhat-marketplace-vtdbz\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.154798 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.598119 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerStarted","Data":"e8b7bfbc3673b348f3db3f6f5913580dacb4ea70e59004aed415c45c038aefd7"} Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.619180 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7664c" podStartSLOduration=3.121475511 podStartE2EDuration="8.619165012s" podCreationTimestamp="2025-12-03 19:36:35 +0000 UTC" firstStartedPulling="2025-12-03 19:36:37.540951271 +0000 UTC m=+9652.742328132" lastFinishedPulling="2025-12-03 19:36:43.038640772 +0000 UTC m=+9658.240017633" observedRunningTime="2025-12-03 19:36:43.61461766 +0000 UTC m=+9658.815994521" watchObservedRunningTime="2025-12-03 19:36:43.619165012 +0000 UTC m=+9658.820541873" Dec 03 19:36:43 crc kubenswrapper[4758]: I1203 19:36:43.636984 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtdbz"] Dec 03 19:36:44 crc kubenswrapper[4758]: I1203 19:36:44.608308 4758 generic.go:334] "Generic (PLEG): container finished" podID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerID="204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83" exitCode=0 Dec 03 19:36:44 crc kubenswrapper[4758]: I1203 19:36:44.608542 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerDied","Data":"204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83"} Dec 03 19:36:44 crc kubenswrapper[4758]: I1203 19:36:44.608614 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerStarted","Data":"3d2b7b3db7bc9e441192e2cb8b9d55397feed8399fefcbb55bfa3f955e9d9c94"} Dec 03 19:36:45 crc kubenswrapper[4758]: I1203 19:36:45.445194 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:45 crc kubenswrapper[4758]: I1203 19:36:45.445732 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:45 crc kubenswrapper[4758]: I1203 19:36:45.621301 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerStarted","Data":"4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1"} Dec 03 19:36:46 crc kubenswrapper[4758]: I1203 19:36:46.526762 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7664c" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="registry-server" probeResult="failure" output=< Dec 03 19:36:46 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 19:36:46 crc kubenswrapper[4758]: > Dec 03 19:36:47 crc kubenswrapper[4758]: I1203 19:36:47.642490 4758 generic.go:334] "Generic (PLEG): container finished" podID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerID="4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1" exitCode=0 Dec 03 19:36:47 crc kubenswrapper[4758]: I1203 19:36:47.642543 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerDied","Data":"4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1"} Dec 03 19:36:48 crc kubenswrapper[4758]: I1203 19:36:48.659411 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerStarted","Data":"8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149"} Dec 03 19:36:48 crc kubenswrapper[4758]: I1203 19:36:48.691228 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vtdbz" podStartSLOduration=3.2329744209999998 podStartE2EDuration="6.691209798s" podCreationTimestamp="2025-12-03 19:36:42 +0000 UTC" firstStartedPulling="2025-12-03 19:36:44.610827718 +0000 UTC m=+9659.812204579" lastFinishedPulling="2025-12-03 19:36:48.069063095 +0000 UTC m=+9663.270439956" observedRunningTime="2025-12-03 19:36:48.690029616 +0000 UTC m=+9663.891406477" watchObservedRunningTime="2025-12-03 19:36:48.691209798 +0000 UTC m=+9663.892586659" Dec 03 19:36:53 crc kubenswrapper[4758]: I1203 19:36:53.155253 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:53 crc kubenswrapper[4758]: I1203 19:36:53.155794 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:53 crc kubenswrapper[4758]: I1203 19:36:53.785526 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:53 crc kubenswrapper[4758]: I1203 19:36:53.839474 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:54 crc kubenswrapper[4758]: I1203 19:36:54.613188 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtdbz"] Dec 03 19:36:55 crc kubenswrapper[4758]: I1203 19:36:55.682009 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:55 crc kubenswrapper[4758]: I1203 19:36:55.734382 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:55 crc kubenswrapper[4758]: I1203 19:36:55.775599 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vtdbz" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="registry-server" containerID="cri-o://8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149" gracePeriod=2 Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.423407 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.509497 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-utilities\") pod \"f05e6807-d0c3-467c-862a-ca31b573f8ab\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.510695 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd7tr\" (UniqueName: \"kubernetes.io/projected/f05e6807-d0c3-467c-862a-ca31b573f8ab-kube-api-access-cd7tr\") pod \"f05e6807-d0c3-467c-862a-ca31b573f8ab\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.511727 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-catalog-content\") pod \"f05e6807-d0c3-467c-862a-ca31b573f8ab\" (UID: \"f05e6807-d0c3-467c-862a-ca31b573f8ab\") " Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.510553 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-utilities" (OuterVolumeSpecName: "utilities") pod "f05e6807-d0c3-467c-862a-ca31b573f8ab" (UID: "f05e6807-d0c3-467c-862a-ca31b573f8ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.516287 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05e6807-d0c3-467c-862a-ca31b573f8ab-kube-api-access-cd7tr" (OuterVolumeSpecName: "kube-api-access-cd7tr") pod "f05e6807-d0c3-467c-862a-ca31b573f8ab" (UID: "f05e6807-d0c3-467c-862a-ca31b573f8ab"). InnerVolumeSpecName "kube-api-access-cd7tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.537735 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f05e6807-d0c3-467c-862a-ca31b573f8ab" (UID: "f05e6807-d0c3-467c-862a-ca31b573f8ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.614764 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd7tr\" (UniqueName: \"kubernetes.io/projected/f05e6807-d0c3-467c-862a-ca31b573f8ab-kube-api-access-cd7tr\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.614817 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.614830 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f05e6807-d0c3-467c-862a-ca31b573f8ab-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.785626 4758 generic.go:334] "Generic (PLEG): container finished" podID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerID="8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149" exitCode=0 Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.785667 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerDied","Data":"8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149"} Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.785698 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtdbz" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.785716 4758 scope.go:117] "RemoveContainer" containerID="8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.785702 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtdbz" event={"ID":"f05e6807-d0c3-467c-862a-ca31b573f8ab","Type":"ContainerDied","Data":"3d2b7b3db7bc9e441192e2cb8b9d55397feed8399fefcbb55bfa3f955e9d9c94"} Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.821559 4758 scope.go:117] "RemoveContainer" containerID="4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.843441 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtdbz"] Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.854771 4758 scope.go:117] "RemoveContainer" containerID="204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.855828 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtdbz"] Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.902791 4758 scope.go:117] "RemoveContainer" containerID="8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149" Dec 03 19:36:56 crc kubenswrapper[4758]: E1203 19:36:56.903279 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149\": container with ID starting with 8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149 not found: ID does not exist" containerID="8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.903319 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149"} err="failed to get container status \"8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149\": rpc error: code = NotFound desc = could not find container \"8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149\": container with ID starting with 8701ed6684c2673bd8c67237896be400b6bd3e70b96c5d2ec1c09de55fd60149 not found: ID does not exist" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.903345 4758 scope.go:117] "RemoveContainer" containerID="4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1" Dec 03 19:36:56 crc kubenswrapper[4758]: E1203 19:36:56.903835 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1\": container with ID starting with 4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1 not found: ID does not exist" containerID="4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.903877 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1"} err="failed to get container status \"4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1\": rpc error: code = NotFound desc = could not find container \"4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1\": container with ID starting with 4e6257d38e910474811429e43eeee48a531c11315d0b6a61b0b6920748c965a1 not found: ID does not exist" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.903904 4758 scope.go:117] "RemoveContainer" containerID="204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83" Dec 03 19:36:56 crc kubenswrapper[4758]: E1203 19:36:56.904182 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83\": container with ID starting with 204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83 not found: ID does not exist" containerID="204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83" Dec 03 19:36:56 crc kubenswrapper[4758]: I1203 19:36:56.904224 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83"} err="failed to get container status \"204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83\": rpc error: code = NotFound desc = could not find container \"204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83\": container with ID starting with 204572edf16056982d94ad6b3427bde54976a5366b3c61a82df0d1a30bb52d83 not found: ID does not exist" Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.126970 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" path="/var/lib/kubelet/pods/f05e6807-d0c3-467c-862a-ca31b573f8ab/volumes" Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.408289 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7664c"] Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.408832 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7664c" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="registry-server" containerID="cri-o://e8b7bfbc3673b348f3db3f6f5913580dacb4ea70e59004aed415c45c038aefd7" gracePeriod=2 Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.822472 4758 generic.go:334] "Generic (PLEG): container finished" podID="91673073-3678-461e-8ed3-24b827f2a252" containerID="e8b7bfbc3673b348f3db3f6f5913580dacb4ea70e59004aed415c45c038aefd7" exitCode=0 Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.822533 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerDied","Data":"e8b7bfbc3673b348f3db3f6f5913580dacb4ea70e59004aed415c45c038aefd7"} Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.822567 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7664c" event={"ID":"91673073-3678-461e-8ed3-24b827f2a252","Type":"ContainerDied","Data":"270cde1f5d733fdc2e570c6e25c184891bf28e9f47a83b5df9238b6462d00705"} Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.822583 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="270cde1f5d733fdc2e570c6e25c184891bf28e9f47a83b5df9238b6462d00705" Dec 03 19:36:57 crc kubenswrapper[4758]: I1203 19:36:57.902611 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.043281 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbhsx\" (UniqueName: \"kubernetes.io/projected/91673073-3678-461e-8ed3-24b827f2a252-kube-api-access-jbhsx\") pod \"91673073-3678-461e-8ed3-24b827f2a252\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.043337 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-catalog-content\") pod \"91673073-3678-461e-8ed3-24b827f2a252\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.043469 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-utilities\") pod \"91673073-3678-461e-8ed3-24b827f2a252\" (UID: \"91673073-3678-461e-8ed3-24b827f2a252\") " Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.044758 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-utilities" (OuterVolumeSpecName: "utilities") pod "91673073-3678-461e-8ed3-24b827f2a252" (UID: "91673073-3678-461e-8ed3-24b827f2a252"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.051926 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91673073-3678-461e-8ed3-24b827f2a252-kube-api-access-jbhsx" (OuterVolumeSpecName: "kube-api-access-jbhsx") pod "91673073-3678-461e-8ed3-24b827f2a252" (UID: "91673073-3678-461e-8ed3-24b827f2a252"). InnerVolumeSpecName "kube-api-access-jbhsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.145762 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbhsx\" (UniqueName: \"kubernetes.io/projected/91673073-3678-461e-8ed3-24b827f2a252-kube-api-access-jbhsx\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.145806 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.155116 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91673073-3678-461e-8ed3-24b827f2a252" (UID: "91673073-3678-461e-8ed3-24b827f2a252"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.247998 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91673073-3678-461e-8ed3-24b827f2a252-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.833752 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7664c" Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.873441 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7664c"] Dec 03 19:36:58 crc kubenswrapper[4758]: I1203 19:36:58.891625 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7664c"] Dec 03 19:36:59 crc kubenswrapper[4758]: I1203 19:36:59.126620 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91673073-3678-461e-8ed3-24b827f2a252" path="/var/lib/kubelet/pods/91673073-3678-461e-8ed3-24b827f2a252/volumes" Dec 03 19:38:41 crc kubenswrapper[4758]: I1203 19:38:41.394319 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:38:41 crc kubenswrapper[4758]: I1203 19:38:41.394891 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:38:42 crc kubenswrapper[4758]: I1203 19:38:42.538993 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 03 19:38:42 crc kubenswrapper[4758]: I1203 19:38:42.539554 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" containerName="adoption" containerID="cri-o://befb8fff4e1f597d0652fa0d78402245fbbbd8e53e64add401f1a71f8515d9dd" gracePeriod=30 Dec 03 19:39:11 crc kubenswrapper[4758]: I1203 19:39:11.394531 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:39:11 crc kubenswrapper[4758]: I1203 19:39:11.395251 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:39:12 crc kubenswrapper[4758]: I1203 19:39:12.770016 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" containerID="befb8fff4e1f597d0652fa0d78402245fbbbd8e53e64add401f1a71f8515d9dd" exitCode=137 Dec 03 19:39:12 crc kubenswrapper[4758]: I1203 19:39:12.770108 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c","Type":"ContainerDied","Data":"befb8fff4e1f597d0652fa0d78402245fbbbd8e53e64add401f1a71f8515d9dd"} Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.062447 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.189980 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvq6h\" (UniqueName: \"kubernetes.io/projected/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c-kube-api-access-lvq6h\") pod \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.191355 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") pod \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\" (UID: \"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c\") " Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.196303 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c-kube-api-access-lvq6h" (OuterVolumeSpecName: "kube-api-access-lvq6h") pod "5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" (UID: "5b166c4c-0f3a-49b8-9a9f-0b1461ead18c"). InnerVolumeSpecName "kube-api-access-lvq6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.218211 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7" (OuterVolumeSpecName: "mariadb-data") pod "5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" (UID: "5b166c4c-0f3a-49b8-9a9f-0b1461ead18c"). InnerVolumeSpecName "pvc-e00c4833-294a-47f9-9733-cdb516be49b7". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.296337 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvq6h\" (UniqueName: \"kubernetes.io/projected/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c-kube-api-access-lvq6h\") on node \"crc\" DevicePath \"\"" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.296392 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e00c4833-294a-47f9-9733-cdb516be49b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") on node \"crc\" " Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.325763 4758 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.325958 4758 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e00c4833-294a-47f9-9733-cdb516be49b7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7") on node "crc" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.398204 4758 reconciler_common.go:293] "Volume detached for volume \"pvc-e00c4833-294a-47f9-9733-cdb516be49b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e00c4833-294a-47f9-9733-cdb516be49b7\") on node \"crc\" DevicePath \"\"" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.782888 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"5b166c4c-0f3a-49b8-9a9f-0b1461ead18c","Type":"ContainerDied","Data":"f263d2e57e84a74476dae8318e6d6ec2941922f0f757e4ecd7117187ee23bb3e"} Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.782946 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.783259 4758 scope.go:117] "RemoveContainer" containerID="befb8fff4e1f597d0652fa0d78402245fbbbd8e53e64add401f1a71f8515d9dd" Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.848471 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 03 19:39:13 crc kubenswrapper[4758]: I1203 19:39:13.859028 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 03 19:39:14 crc kubenswrapper[4758]: I1203 19:39:14.630914 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 03 19:39:14 crc kubenswrapper[4758]: I1203 19:39:14.631350 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="3206afc1-0965-4c69-a90c-95be3a0605b3" containerName="adoption" containerID="cri-o://37a9efb4e5ef7c6a7fb36e6cf8cba641b6971768e8011861d63f186d28ee2af3" gracePeriod=30 Dec 03 19:39:15 crc kubenswrapper[4758]: I1203 19:39:15.126469 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" path="/var/lib/kubelet/pods/5b166c4c-0f3a-49b8-9a9f-0b1461ead18c/volumes" Dec 03 19:39:41 crc kubenswrapper[4758]: I1203 19:39:41.394262 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:39:41 crc kubenswrapper[4758]: I1203 19:39:41.394847 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:39:41 crc kubenswrapper[4758]: I1203 19:39:41.394893 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:39:41 crc kubenswrapper[4758]: I1203 19:39:41.395737 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21a90d7bea8cfab67e8c91f5063700f03e750127eee664fa2caca4a804f6cffa"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:39:41 crc kubenswrapper[4758]: I1203 19:39:41.395792 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://21a90d7bea8cfab67e8c91f5063700f03e750127eee664fa2caca4a804f6cffa" gracePeriod=600 Dec 03 19:39:42 crc kubenswrapper[4758]: I1203 19:39:42.068373 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="21a90d7bea8cfab67e8c91f5063700f03e750127eee664fa2caca4a804f6cffa" exitCode=0 Dec 03 19:39:42 crc kubenswrapper[4758]: I1203 19:39:42.068424 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"21a90d7bea8cfab67e8c91f5063700f03e750127eee664fa2caca4a804f6cffa"} Dec 03 19:39:42 crc kubenswrapper[4758]: I1203 19:39:42.068765 4758 scope.go:117] "RemoveContainer" containerID="c58060ec7316231d6b6de1dac56a50a7ce6da0b9eb7354a9a2505fb748d021ce" Dec 03 19:39:43 crc kubenswrapper[4758]: I1203 19:39:43.082559 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727"} Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.105312 4758 generic.go:334] "Generic (PLEG): container finished" podID="3206afc1-0965-4c69-a90c-95be3a0605b3" containerID="37a9efb4e5ef7c6a7fb36e6cf8cba641b6971768e8011861d63f186d28ee2af3" exitCode=137 Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.105407 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3206afc1-0965-4c69-a90c-95be3a0605b3","Type":"ContainerDied","Data":"37a9efb4e5ef7c6a7fb36e6cf8cba641b6971768e8011861d63f186d28ee2af3"} Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.688903 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.780199 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3206afc1-0965-4c69-a90c-95be3a0605b3-ovn-data-cert\") pod \"3206afc1-0965-4c69-a90c-95be3a0605b3\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.780383 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnxb2\" (UniqueName: \"kubernetes.io/projected/3206afc1-0965-4c69-a90c-95be3a0605b3-kube-api-access-pnxb2\") pod \"3206afc1-0965-4c69-a90c-95be3a0605b3\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.781137 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") pod \"3206afc1-0965-4c69-a90c-95be3a0605b3\" (UID: \"3206afc1-0965-4c69-a90c-95be3a0605b3\") " Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.785809 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3206afc1-0965-4c69-a90c-95be3a0605b3-kube-api-access-pnxb2" (OuterVolumeSpecName: "kube-api-access-pnxb2") pod "3206afc1-0965-4c69-a90c-95be3a0605b3" (UID: "3206afc1-0965-4c69-a90c-95be3a0605b3"). InnerVolumeSpecName "kube-api-access-pnxb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.785893 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3206afc1-0965-4c69-a90c-95be3a0605b3-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "3206afc1-0965-4c69-a90c-95be3a0605b3" (UID: "3206afc1-0965-4c69-a90c-95be3a0605b3"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.803743 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0" (OuterVolumeSpecName: "ovn-data") pod "3206afc1-0965-4c69-a90c-95be3a0605b3" (UID: "3206afc1-0965-4c69-a90c-95be3a0605b3"). InnerVolumeSpecName "pvc-5baafff9-4b8b-4714-8758-87fa113058a0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.884391 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5baafff9-4b8b-4714-8758-87fa113058a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") on node \"crc\" " Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.884440 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3206afc1-0965-4c69-a90c-95be3a0605b3-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.884455 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnxb2\" (UniqueName: \"kubernetes.io/projected/3206afc1-0965-4c69-a90c-95be3a0605b3-kube-api-access-pnxb2\") on node \"crc\" DevicePath \"\"" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.921462 4758 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.921672 4758 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5baafff9-4b8b-4714-8758-87fa113058a0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0") on node "crc" Dec 03 19:39:45 crc kubenswrapper[4758]: I1203 19:39:45.986085 4758 reconciler_common.go:293] "Volume detached for volume \"pvc-5baafff9-4b8b-4714-8758-87fa113058a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5baafff9-4b8b-4714-8758-87fa113058a0\") on node \"crc\" DevicePath \"\"" Dec 03 19:39:46 crc kubenswrapper[4758]: I1203 19:39:46.117806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3206afc1-0965-4c69-a90c-95be3a0605b3","Type":"ContainerDied","Data":"13912022e8ffef461395dfc9683b99e8d48dc59178c2a16f929d00ce3cb93a5a"} Dec 03 19:39:46 crc kubenswrapper[4758]: I1203 19:39:46.117850 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 03 19:39:46 crc kubenswrapper[4758]: I1203 19:39:46.118218 4758 scope.go:117] "RemoveContainer" containerID="37a9efb4e5ef7c6a7fb36e6cf8cba641b6971768e8011861d63f186d28ee2af3" Dec 03 19:39:46 crc kubenswrapper[4758]: I1203 19:39:46.215778 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 03 19:39:46 crc kubenswrapper[4758]: I1203 19:39:46.271520 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 03 19:39:47 crc kubenswrapper[4758]: I1203 19:39:47.129670 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3206afc1-0965-4c69-a90c-95be3a0605b3" path="/var/lib/kubelet/pods/3206afc1-0965-4c69-a90c-95be3a0605b3/volumes" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.861656 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s29l4/must-gather-48wk6"] Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862832 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" containerName="adoption" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.862849 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" containerName="adoption" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862873 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="registry-server" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.862880 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="registry-server" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862893 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="extract-utilities" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.862902 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="extract-utilities" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862925 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="extract-utilities" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.862934 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="extract-utilities" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862945 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3206afc1-0965-4c69-a90c-95be3a0605b3" containerName="adoption" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.862952 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3206afc1-0965-4c69-a90c-95be3a0605b3" containerName="adoption" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862973 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="extract-content" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.862981 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="extract-content" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.862993 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="registry-server" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.863000 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="registry-server" Dec 03 19:40:43 crc kubenswrapper[4758]: E1203 19:40:43.863030 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="extract-content" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.863037 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="extract-content" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.863300 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="91673073-3678-461e-8ed3-24b827f2a252" containerName="registry-server" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.863319 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b166c4c-0f3a-49b8-9a9f-0b1461ead18c" containerName="adoption" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.863340 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3206afc1-0965-4c69-a90c-95be3a0605b3" containerName="adoption" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.863356 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f05e6807-d0c3-467c-862a-ca31b573f8ab" containerName="registry-server" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.864857 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.870329 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s29l4"/"openshift-service-ca.crt" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.870329 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-s29l4"/"default-dockercfg-8n2gb" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.872031 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s29l4"/"kube-root-ca.crt" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.884636 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s29l4/must-gather-48wk6"] Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.973797 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cb2432eb-1022-4475-ba37-c91e5435f875-must-gather-output\") pod \"must-gather-48wk6\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:43 crc kubenswrapper[4758]: I1203 19:40:43.974255 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl2zx\" (UniqueName: \"kubernetes.io/projected/cb2432eb-1022-4475-ba37-c91e5435f875-kube-api-access-nl2zx\") pod \"must-gather-48wk6\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.077172 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl2zx\" (UniqueName: \"kubernetes.io/projected/cb2432eb-1022-4475-ba37-c91e5435f875-kube-api-access-nl2zx\") pod \"must-gather-48wk6\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.077670 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cb2432eb-1022-4475-ba37-c91e5435f875-must-gather-output\") pod \"must-gather-48wk6\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.078188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cb2432eb-1022-4475-ba37-c91e5435f875-must-gather-output\") pod \"must-gather-48wk6\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.100186 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl2zx\" (UniqueName: \"kubernetes.io/projected/cb2432eb-1022-4475-ba37-c91e5435f875-kube-api-access-nl2zx\") pod \"must-gather-48wk6\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.195610 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.676109 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s29l4/must-gather-48wk6"] Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.681596 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:40:44 crc kubenswrapper[4758]: I1203 19:40:44.718229 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/must-gather-48wk6" event={"ID":"cb2432eb-1022-4475-ba37-c91e5435f875","Type":"ContainerStarted","Data":"1277f4a6eaad6cb3ae2aefad9669275b4957537a38df2444fe5efa814882db5e"} Dec 03 19:40:49 crc kubenswrapper[4758]: I1203 19:40:49.798096 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/must-gather-48wk6" event={"ID":"cb2432eb-1022-4475-ba37-c91e5435f875","Type":"ContainerStarted","Data":"367f6f1f74a93694c8e369ae18ec950d5c2e0c779556c72786652f642eb02cc4"} Dec 03 19:40:50 crc kubenswrapper[4758]: I1203 19:40:50.818250 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/must-gather-48wk6" event={"ID":"cb2432eb-1022-4475-ba37-c91e5435f875","Type":"ContainerStarted","Data":"133d6a89ec23246e89daa027b2760c2afadcf1e81c285f7346e3764a365a7d37"} Dec 03 19:40:50 crc kubenswrapper[4758]: I1203 19:40:50.846911 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s29l4/must-gather-48wk6" podStartSLOduration=3.382597576 podStartE2EDuration="7.846889354s" podCreationTimestamp="2025-12-03 19:40:43 +0000 UTC" firstStartedPulling="2025-12-03 19:40:44.681314118 +0000 UTC m=+9899.882690979" lastFinishedPulling="2025-12-03 19:40:49.145605896 +0000 UTC m=+9904.346982757" observedRunningTime="2025-12-03 19:40:50.838273294 +0000 UTC m=+9906.039650165" watchObservedRunningTime="2025-12-03 19:40:50.846889354 +0000 UTC m=+9906.048266225" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.052602 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s29l4/crc-debug-vg4gp"] Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.056332 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.226715 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4vrk\" (UniqueName: \"kubernetes.io/projected/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-kube-api-access-g4vrk\") pod \"crc-debug-vg4gp\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.227207 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-host\") pod \"crc-debug-vg4gp\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.329019 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4vrk\" (UniqueName: \"kubernetes.io/projected/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-kube-api-access-g4vrk\") pod \"crc-debug-vg4gp\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.329108 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-host\") pod \"crc-debug-vg4gp\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.329196 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-host\") pod \"crc-debug-vg4gp\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.363347 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4vrk\" (UniqueName: \"kubernetes.io/projected/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-kube-api-access-g4vrk\") pod \"crc-debug-vg4gp\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.380562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:40:54 crc kubenswrapper[4758]: I1203 19:40:54.859169 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" event={"ID":"0cfaa103-d889-4a39-b0d5-5c7794e12ff1","Type":"ContainerStarted","Data":"a3262d8a0d15c7e55278ec3b0c9db69a4947827af368e0fc9d5c011f943af229"} Dec 03 19:41:11 crc kubenswrapper[4758]: I1203 19:41:11.076365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" event={"ID":"0cfaa103-d889-4a39-b0d5-5c7794e12ff1","Type":"ContainerStarted","Data":"7f4f463f05ef45138fdd84da78fd57eb36efcffd72d20498b7a09d206a75a2c5"} Dec 03 19:41:11 crc kubenswrapper[4758]: I1203 19:41:11.100756 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" podStartSLOduration=1.440662339 podStartE2EDuration="17.100630548s" podCreationTimestamp="2025-12-03 19:40:54 +0000 UTC" firstStartedPulling="2025-12-03 19:40:54.43557017 +0000 UTC m=+9909.636947031" lastFinishedPulling="2025-12-03 19:41:10.095538379 +0000 UTC m=+9925.296915240" observedRunningTime="2025-12-03 19:41:11.090097916 +0000 UTC m=+9926.291474777" watchObservedRunningTime="2025-12-03 19:41:11.100630548 +0000 UTC m=+9926.302007409" Dec 03 19:41:19 crc kubenswrapper[4758]: I1203 19:41:19.907391 4758 scope.go:117] "RemoveContainer" containerID="60e4bbf36048bf623888201548c0e5934990eafab202fcfed8194eb175b4b891" Dec 03 19:41:21 crc kubenswrapper[4758]: I1203 19:41:21.537384 4758 scope.go:117] "RemoveContainer" containerID="b22764317f4e2908089e4ee776a4ca76be39d76865234bf44c82f81c5adf7464" Dec 03 19:41:22 crc kubenswrapper[4758]: I1203 19:41:22.316229 4758 scope.go:117] "RemoveContainer" containerID="a837e1c8507ac52f2469f2392917c0682c712b1d3f8b02984ad8838c89fe8163" Dec 03 19:41:32 crc kubenswrapper[4758]: I1203 19:41:32.347282 4758 generic.go:334] "Generic (PLEG): container finished" podID="0cfaa103-d889-4a39-b0d5-5c7794e12ff1" containerID="7f4f463f05ef45138fdd84da78fd57eb36efcffd72d20498b7a09d206a75a2c5" exitCode=0 Dec 03 19:41:32 crc kubenswrapper[4758]: I1203 19:41:32.347795 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" event={"ID":"0cfaa103-d889-4a39-b0d5-5c7794e12ff1","Type":"ContainerDied","Data":"7f4f463f05ef45138fdd84da78fd57eb36efcffd72d20498b7a09d206a75a2c5"} Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.651356 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.695361 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s29l4/crc-debug-vg4gp"] Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.708785 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s29l4/crc-debug-vg4gp"] Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.756585 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-host\") pod \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.756754 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-host" (OuterVolumeSpecName: "host") pod "0cfaa103-d889-4a39-b0d5-5c7794e12ff1" (UID: "0cfaa103-d889-4a39-b0d5-5c7794e12ff1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.756850 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4vrk\" (UniqueName: \"kubernetes.io/projected/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-kube-api-access-g4vrk\") pod \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\" (UID: \"0cfaa103-d889-4a39-b0d5-5c7794e12ff1\") " Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.757361 4758 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.763936 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-kube-api-access-g4vrk" (OuterVolumeSpecName: "kube-api-access-g4vrk") pod "0cfaa103-d889-4a39-b0d5-5c7794e12ff1" (UID: "0cfaa103-d889-4a39-b0d5-5c7794e12ff1"). InnerVolumeSpecName "kube-api-access-g4vrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:41:33 crc kubenswrapper[4758]: I1203 19:41:33.859592 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4vrk\" (UniqueName: \"kubernetes.io/projected/0cfaa103-d889-4a39-b0d5-5c7794e12ff1-kube-api-access-g4vrk\") on node \"crc\" DevicePath \"\"" Dec 03 19:41:34 crc kubenswrapper[4758]: I1203 19:41:34.368712 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3262d8a0d15c7e55278ec3b0c9db69a4947827af368e0fc9d5c011f943af229" Dec 03 19:41:34 crc kubenswrapper[4758]: I1203 19:41:34.368806 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-vg4gp" Dec 03 19:41:34 crc kubenswrapper[4758]: I1203 19:41:34.968184 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s29l4/crc-debug-zjdxh"] Dec 03 19:41:34 crc kubenswrapper[4758]: E1203 19:41:34.969501 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfaa103-d889-4a39-b0d5-5c7794e12ff1" containerName="container-00" Dec 03 19:41:34 crc kubenswrapper[4758]: I1203 19:41:34.969516 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfaa103-d889-4a39-b0d5-5c7794e12ff1" containerName="container-00" Dec 03 19:41:34 crc kubenswrapper[4758]: I1203 19:41:34.969885 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfaa103-d889-4a39-b0d5-5c7794e12ff1" containerName="container-00" Dec 03 19:41:34 crc kubenswrapper[4758]: I1203 19:41:34.973056 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.002077 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74848554-5483-49be-b76b-0df0859b288b-host\") pod \"crc-debug-zjdxh\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.002167 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mxrh\" (UniqueName: \"kubernetes.io/projected/74848554-5483-49be-b76b-0df0859b288b-kube-api-access-9mxrh\") pod \"crc-debug-zjdxh\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.104627 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74848554-5483-49be-b76b-0df0859b288b-host\") pod \"crc-debug-zjdxh\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.104734 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mxrh\" (UniqueName: \"kubernetes.io/projected/74848554-5483-49be-b76b-0df0859b288b-kube-api-access-9mxrh\") pod \"crc-debug-zjdxh\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.104942 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74848554-5483-49be-b76b-0df0859b288b-host\") pod \"crc-debug-zjdxh\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.133821 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfaa103-d889-4a39-b0d5-5c7794e12ff1" path="/var/lib/kubelet/pods/0cfaa103-d889-4a39-b0d5-5c7794e12ff1/volumes" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.134426 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mxrh\" (UniqueName: \"kubernetes.io/projected/74848554-5483-49be-b76b-0df0859b288b-kube-api-access-9mxrh\") pod \"crc-debug-zjdxh\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:35 crc kubenswrapper[4758]: I1203 19:41:35.300456 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:36 crc kubenswrapper[4758]: I1203 19:41:36.393488 4758 generic.go:334] "Generic (PLEG): container finished" podID="74848554-5483-49be-b76b-0df0859b288b" containerID="2b5ab7f0e62b327490fc15a60ab98fa7316bc368a585e0263e6ff98c348f7aa6" exitCode=1 Dec 03 19:41:36 crc kubenswrapper[4758]: I1203 19:41:36.393582 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/crc-debug-zjdxh" event={"ID":"74848554-5483-49be-b76b-0df0859b288b","Type":"ContainerDied","Data":"2b5ab7f0e62b327490fc15a60ab98fa7316bc368a585e0263e6ff98c348f7aa6"} Dec 03 19:41:36 crc kubenswrapper[4758]: I1203 19:41:36.393940 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/crc-debug-zjdxh" event={"ID":"74848554-5483-49be-b76b-0df0859b288b","Type":"ContainerStarted","Data":"78b1457a18ea16acafec7a4383dd97a7421bf4ad02758beb68667d6ed66f192a"} Dec 03 19:41:36 crc kubenswrapper[4758]: I1203 19:41:36.448722 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s29l4/crc-debug-zjdxh"] Dec 03 19:41:36 crc kubenswrapper[4758]: I1203 19:41:36.462971 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s29l4/crc-debug-zjdxh"] Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.538157 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.675471 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74848554-5483-49be-b76b-0df0859b288b-host\") pod \"74848554-5483-49be-b76b-0df0859b288b\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.675619 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74848554-5483-49be-b76b-0df0859b288b-host" (OuterVolumeSpecName: "host") pod "74848554-5483-49be-b76b-0df0859b288b" (UID: "74848554-5483-49be-b76b-0df0859b288b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.675739 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mxrh\" (UniqueName: \"kubernetes.io/projected/74848554-5483-49be-b76b-0df0859b288b-kube-api-access-9mxrh\") pod \"74848554-5483-49be-b76b-0df0859b288b\" (UID: \"74848554-5483-49be-b76b-0df0859b288b\") " Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.676454 4758 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74848554-5483-49be-b76b-0df0859b288b-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.685960 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74848554-5483-49be-b76b-0df0859b288b-kube-api-access-9mxrh" (OuterVolumeSpecName: "kube-api-access-9mxrh") pod "74848554-5483-49be-b76b-0df0859b288b" (UID: "74848554-5483-49be-b76b-0df0859b288b"). InnerVolumeSpecName "kube-api-access-9mxrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:41:37 crc kubenswrapper[4758]: I1203 19:41:37.778378 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mxrh\" (UniqueName: \"kubernetes.io/projected/74848554-5483-49be-b76b-0df0859b288b-kube-api-access-9mxrh\") on node \"crc\" DevicePath \"\"" Dec 03 19:41:38 crc kubenswrapper[4758]: I1203 19:41:38.444608 4758 scope.go:117] "RemoveContainer" containerID="2b5ab7f0e62b327490fc15a60ab98fa7316bc368a585e0263e6ff98c348f7aa6" Dec 03 19:41:38 crc kubenswrapper[4758]: I1203 19:41:38.444712 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/crc-debug-zjdxh" Dec 03 19:41:39 crc kubenswrapper[4758]: I1203 19:41:39.127888 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74848554-5483-49be-b76b-0df0859b288b" path="/var/lib/kubelet/pods/74848554-5483-49be-b76b-0df0859b288b/volumes" Dec 03 19:42:11 crc kubenswrapper[4758]: I1203 19:42:11.394486 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:42:11 crc kubenswrapper[4758]: I1203 19:42:11.395115 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:42:41 crc kubenswrapper[4758]: I1203 19:42:41.394187 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:42:41 crc kubenswrapper[4758]: I1203 19:42:41.394647 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:43:11 crc kubenswrapper[4758]: I1203 19:43:11.395210 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:43:11 crc kubenswrapper[4758]: I1203 19:43:11.396361 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:43:11 crc kubenswrapper[4758]: I1203 19:43:11.396448 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:43:11 crc kubenswrapper[4758]: I1203 19:43:11.398069 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:43:11 crc kubenswrapper[4758]: I1203 19:43:11.398157 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" gracePeriod=600 Dec 03 19:43:11 crc kubenswrapper[4758]: E1203 19:43:11.537290 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:43:12 crc kubenswrapper[4758]: I1203 19:43:12.441160 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" exitCode=0 Dec 03 19:43:12 crc kubenswrapper[4758]: I1203 19:43:12.441260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727"} Dec 03 19:43:12 crc kubenswrapper[4758]: I1203 19:43:12.441899 4758 scope.go:117] "RemoveContainer" containerID="21a90d7bea8cfab67e8c91f5063700f03e750127eee664fa2caca4a804f6cffa" Dec 03 19:43:12 crc kubenswrapper[4758]: I1203 19:43:12.442627 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:43:12 crc kubenswrapper[4758]: E1203 19:43:12.442934 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:43:22 crc kubenswrapper[4758]: I1203 19:43:22.495592 4758 scope.go:117] "RemoveContainer" containerID="e8b7bfbc3673b348f3db3f6f5913580dacb4ea70e59004aed415c45c038aefd7" Dec 03 19:43:22 crc kubenswrapper[4758]: I1203 19:43:22.521832 4758 scope.go:117] "RemoveContainer" containerID="22287e750fc817267c3d5867b06d5e74ac4708e494535b3364c9ae07754ac877" Dec 03 19:43:22 crc kubenswrapper[4758]: I1203 19:43:22.556952 4758 scope.go:117] "RemoveContainer" containerID="cda70aefffa7c97e2d75c7b628b8fadd2748247d6eb6d12f0e40e003ae526307" Dec 03 19:43:27 crc kubenswrapper[4758]: I1203 19:43:27.114470 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:43:27 crc kubenswrapper[4758]: E1203 19:43:27.115087 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:43:42 crc kubenswrapper[4758]: I1203 19:43:42.115405 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:43:42 crc kubenswrapper[4758]: E1203 19:43:42.116526 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:43:56 crc kubenswrapper[4758]: I1203 19:43:56.113878 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:43:56 crc kubenswrapper[4758]: E1203 19:43:56.114976 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:44:07 crc kubenswrapper[4758]: I1203 19:44:07.114415 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:44:07 crc kubenswrapper[4758]: E1203 19:44:07.115483 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:44:18 crc kubenswrapper[4758]: I1203 19:44:18.114138 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:44:18 crc kubenswrapper[4758]: E1203 19:44:18.114950 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:44:29 crc kubenswrapper[4758]: I1203 19:44:29.115498 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:44:29 crc kubenswrapper[4758]: E1203 19:44:29.116212 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:44:41 crc kubenswrapper[4758]: I1203 19:44:41.127885 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:44:41 crc kubenswrapper[4758]: E1203 19:44:41.129024 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:44:54 crc kubenswrapper[4758]: I1203 19:44:54.114126 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:44:54 crc kubenswrapper[4758]: E1203 19:44:54.114808 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.151247 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2"] Dec 03 19:45:00 crc kubenswrapper[4758]: E1203 19:45:00.152605 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74848554-5483-49be-b76b-0df0859b288b" containerName="container-00" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.152624 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="74848554-5483-49be-b76b-0df0859b288b" containerName="container-00" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.152866 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="74848554-5483-49be-b76b-0df0859b288b" containerName="container-00" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.153839 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.161222 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.161267 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.176958 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2"] Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.274609 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3046236b-1ddf-491b-a83e-3f8b977fe383-secret-volume\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.274913 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzgl8\" (UniqueName: \"kubernetes.io/projected/3046236b-1ddf-491b-a83e-3f8b977fe383-kube-api-access-gzgl8\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.275403 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3046236b-1ddf-491b-a83e-3f8b977fe383-config-volume\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.377034 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzgl8\" (UniqueName: \"kubernetes.io/projected/3046236b-1ddf-491b-a83e-3f8b977fe383-kube-api-access-gzgl8\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.377251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3046236b-1ddf-491b-a83e-3f8b977fe383-config-volume\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.377372 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3046236b-1ddf-491b-a83e-3f8b977fe383-secret-volume\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.378246 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3046236b-1ddf-491b-a83e-3f8b977fe383-config-volume\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.386506 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3046236b-1ddf-491b-a83e-3f8b977fe383-secret-volume\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.398208 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzgl8\" (UniqueName: \"kubernetes.io/projected/3046236b-1ddf-491b-a83e-3f8b977fe383-kube-api-access-gzgl8\") pod \"collect-profiles-29413185-tn5t2\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:00 crc kubenswrapper[4758]: I1203 19:45:00.493111 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:01 crc kubenswrapper[4758]: I1203 19:45:01.026571 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2"] Dec 03 19:45:01 crc kubenswrapper[4758]: I1203 19:45:01.651722 4758 generic.go:334] "Generic (PLEG): container finished" podID="3046236b-1ddf-491b-a83e-3f8b977fe383" containerID="8ec27b5f8c6db3d2d01fe92c4ccb35238056c22cea3b1c040b2455e9860e0e9f" exitCode=0 Dec 03 19:45:01 crc kubenswrapper[4758]: I1203 19:45:01.651839 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" event={"ID":"3046236b-1ddf-491b-a83e-3f8b977fe383","Type":"ContainerDied","Data":"8ec27b5f8c6db3d2d01fe92c4ccb35238056c22cea3b1c040b2455e9860e0e9f"} Dec 03 19:45:01 crc kubenswrapper[4758]: I1203 19:45:01.652308 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" event={"ID":"3046236b-1ddf-491b-a83e-3f8b977fe383","Type":"ContainerStarted","Data":"a57bab7594e5a20a8d72b64d9bae2613c7b4fe72e44070a6ab9effa8e6d24599"} Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.159092 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.247614 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3046236b-1ddf-491b-a83e-3f8b977fe383-config-volume\") pod \"3046236b-1ddf-491b-a83e-3f8b977fe383\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.247808 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzgl8\" (UniqueName: \"kubernetes.io/projected/3046236b-1ddf-491b-a83e-3f8b977fe383-kube-api-access-gzgl8\") pod \"3046236b-1ddf-491b-a83e-3f8b977fe383\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.248404 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3046236b-1ddf-491b-a83e-3f8b977fe383-secret-volume\") pod \"3046236b-1ddf-491b-a83e-3f8b977fe383\" (UID: \"3046236b-1ddf-491b-a83e-3f8b977fe383\") " Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.248670 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3046236b-1ddf-491b-a83e-3f8b977fe383-config-volume" (OuterVolumeSpecName: "config-volume") pod "3046236b-1ddf-491b-a83e-3f8b977fe383" (UID: "3046236b-1ddf-491b-a83e-3f8b977fe383"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.255925 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3046236b-1ddf-491b-a83e-3f8b977fe383-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.262085 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3046236b-1ddf-491b-a83e-3f8b977fe383-kube-api-access-gzgl8" (OuterVolumeSpecName: "kube-api-access-gzgl8") pod "3046236b-1ddf-491b-a83e-3f8b977fe383" (UID: "3046236b-1ddf-491b-a83e-3f8b977fe383"). InnerVolumeSpecName "kube-api-access-gzgl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.272336 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3046236b-1ddf-491b-a83e-3f8b977fe383-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3046236b-1ddf-491b-a83e-3f8b977fe383" (UID: "3046236b-1ddf-491b-a83e-3f8b977fe383"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.358279 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3046236b-1ddf-491b-a83e-3f8b977fe383-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.358327 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzgl8\" (UniqueName: \"kubernetes.io/projected/3046236b-1ddf-491b-a83e-3f8b977fe383-kube-api-access-gzgl8\") on node \"crc\" DevicePath \"\"" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.695643 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" event={"ID":"3046236b-1ddf-491b-a83e-3f8b977fe383","Type":"ContainerDied","Data":"a57bab7594e5a20a8d72b64d9bae2613c7b4fe72e44070a6ab9effa8e6d24599"} Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.695948 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a57bab7594e5a20a8d72b64d9bae2613c7b4fe72e44070a6ab9effa8e6d24599" Dec 03 19:45:03 crc kubenswrapper[4758]: I1203 19:45:03.695861 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413185-tn5t2" Dec 03 19:45:04 crc kubenswrapper[4758]: I1203 19:45:04.237122 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9"] Dec 03 19:45:04 crc kubenswrapper[4758]: I1203 19:45:04.246428 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-45tx9"] Dec 03 19:45:05 crc kubenswrapper[4758]: I1203 19:45:05.126603 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:45:05 crc kubenswrapper[4758]: E1203 19:45:05.127666 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:45:05 crc kubenswrapper[4758]: I1203 19:45:05.129730 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d7da42-ab46-4ef8-aede-79f3489ed18c" path="/var/lib/kubelet/pods/03d7da42-ab46-4ef8-aede-79f3489ed18c/volumes" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.306818 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nz9xf"] Dec 03 19:45:06 crc kubenswrapper[4758]: E1203 19:45:06.307363 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3046236b-1ddf-491b-a83e-3f8b977fe383" containerName="collect-profiles" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.307380 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3046236b-1ddf-491b-a83e-3f8b977fe383" containerName="collect-profiles" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.307714 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3046236b-1ddf-491b-a83e-3f8b977fe383" containerName="collect-profiles" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.309515 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.323977 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nz9xf"] Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.436502 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-utilities\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.436876 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7wlz\" (UniqueName: \"kubernetes.io/projected/cb1143fc-3616-410b-bb3e-1ba70ad27979-kube-api-access-t7wlz\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.437036 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-catalog-content\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.539835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-utilities\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.539897 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7wlz\" (UniqueName: \"kubernetes.io/projected/cb1143fc-3616-410b-bb3e-1ba70ad27979-kube-api-access-t7wlz\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.539941 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-catalog-content\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.540533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-catalog-content\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.540861 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-utilities\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.572483 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7wlz\" (UniqueName: \"kubernetes.io/projected/cb1143fc-3616-410b-bb3e-1ba70ad27979-kube-api-access-t7wlz\") pod \"certified-operators-nz9xf\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:06 crc kubenswrapper[4758]: I1203 19:45:06.638303 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:07 crc kubenswrapper[4758]: I1203 19:45:07.344635 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nz9xf"] Dec 03 19:45:07 crc kubenswrapper[4758]: I1203 19:45:07.750021 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerID="f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901" exitCode=0 Dec 03 19:45:07 crc kubenswrapper[4758]: I1203 19:45:07.750171 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerDied","Data":"f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901"} Dec 03 19:45:07 crc kubenswrapper[4758]: I1203 19:45:07.750313 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerStarted","Data":"68e048f17793cad4a825aed1770d77d751f4d08b23532c52c712637566397d3d"} Dec 03 19:45:08 crc kubenswrapper[4758]: I1203 19:45:08.762172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerStarted","Data":"0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57"} Dec 03 19:45:10 crc kubenswrapper[4758]: I1203 19:45:10.784463 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerID="0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57" exitCode=0 Dec 03 19:45:10 crc kubenswrapper[4758]: I1203 19:45:10.784541 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerDied","Data":"0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57"} Dec 03 19:45:11 crc kubenswrapper[4758]: I1203 19:45:11.800227 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerStarted","Data":"db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147"} Dec 03 19:45:11 crc kubenswrapper[4758]: I1203 19:45:11.824375 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nz9xf" podStartSLOduration=2.387449485 podStartE2EDuration="5.824357268s" podCreationTimestamp="2025-12-03 19:45:06 +0000 UTC" firstStartedPulling="2025-12-03 19:45:07.752870691 +0000 UTC m=+10162.954247552" lastFinishedPulling="2025-12-03 19:45:11.189778474 +0000 UTC m=+10166.391155335" observedRunningTime="2025-12-03 19:45:11.817503647 +0000 UTC m=+10167.018880508" watchObservedRunningTime="2025-12-03 19:45:11.824357268 +0000 UTC m=+10167.025734129" Dec 03 19:45:16 crc kubenswrapper[4758]: I1203 19:45:16.639019 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:16 crc kubenswrapper[4758]: I1203 19:45:16.639645 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:16 crc kubenswrapper[4758]: I1203 19:45:16.731653 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:16 crc kubenswrapper[4758]: I1203 19:45:16.957074 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:17 crc kubenswrapper[4758]: I1203 19:45:17.017941 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nz9xf"] Dec 03 19:45:18 crc kubenswrapper[4758]: I1203 19:45:18.886851 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nz9xf" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="registry-server" containerID="cri-o://db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147" gracePeriod=2 Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.472105 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.564521 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-catalog-content\") pod \"cb1143fc-3616-410b-bb3e-1ba70ad27979\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.565107 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-utilities\") pod \"cb1143fc-3616-410b-bb3e-1ba70ad27979\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.565130 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7wlz\" (UniqueName: \"kubernetes.io/projected/cb1143fc-3616-410b-bb3e-1ba70ad27979-kube-api-access-t7wlz\") pod \"cb1143fc-3616-410b-bb3e-1ba70ad27979\" (UID: \"cb1143fc-3616-410b-bb3e-1ba70ad27979\") " Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.566638 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-utilities" (OuterVolumeSpecName: "utilities") pod "cb1143fc-3616-410b-bb3e-1ba70ad27979" (UID: "cb1143fc-3616-410b-bb3e-1ba70ad27979"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.574200 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb1143fc-3616-410b-bb3e-1ba70ad27979-kube-api-access-t7wlz" (OuterVolumeSpecName: "kube-api-access-t7wlz") pod "cb1143fc-3616-410b-bb3e-1ba70ad27979" (UID: "cb1143fc-3616-410b-bb3e-1ba70ad27979"). InnerVolumeSpecName "kube-api-access-t7wlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.617581 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb1143fc-3616-410b-bb3e-1ba70ad27979" (UID: "cb1143fc-3616-410b-bb3e-1ba70ad27979"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.668534 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.668578 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1143fc-3616-410b-bb3e-1ba70ad27979-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.668588 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7wlz\" (UniqueName: \"kubernetes.io/projected/cb1143fc-3616-410b-bb3e-1ba70ad27979-kube-api-access-t7wlz\") on node \"crc\" DevicePath \"\"" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.899798 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerID="db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147" exitCode=0 Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.899851 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerDied","Data":"db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147"} Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.899920 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz9xf" event={"ID":"cb1143fc-3616-410b-bb3e-1ba70ad27979","Type":"ContainerDied","Data":"68e048f17793cad4a825aed1770d77d751f4d08b23532c52c712637566397d3d"} Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.899946 4758 scope.go:117] "RemoveContainer" containerID="db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.899877 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz9xf" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.922609 4758 scope.go:117] "RemoveContainer" containerID="0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57" Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.939342 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nz9xf"] Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.950157 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nz9xf"] Dec 03 19:45:19 crc kubenswrapper[4758]: I1203 19:45:19.961050 4758 scope.go:117] "RemoveContainer" containerID="f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.002848 4758 scope.go:117] "RemoveContainer" containerID="db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147" Dec 03 19:45:20 crc kubenswrapper[4758]: E1203 19:45:20.003595 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147\": container with ID starting with db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147 not found: ID does not exist" containerID="db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.003636 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147"} err="failed to get container status \"db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147\": rpc error: code = NotFound desc = could not find container \"db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147\": container with ID starting with db33ec727af6e38b1d8ab9c70f4af9beda8e82384b4a282f7d69128c33cd5147 not found: ID does not exist" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.003669 4758 scope.go:117] "RemoveContainer" containerID="0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57" Dec 03 19:45:20 crc kubenswrapper[4758]: E1203 19:45:20.004236 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57\": container with ID starting with 0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57 not found: ID does not exist" containerID="0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.004322 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57"} err="failed to get container status \"0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57\": rpc error: code = NotFound desc = could not find container \"0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57\": container with ID starting with 0b3722ae659dce024c003a6711e27b12c48eaeb82d1fab4981bc85cec7947d57 not found: ID does not exist" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.004354 4758 scope.go:117] "RemoveContainer" containerID="f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901" Dec 03 19:45:20 crc kubenswrapper[4758]: E1203 19:45:20.004834 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901\": container with ID starting with f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901 not found: ID does not exist" containerID="f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.004888 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901"} err="failed to get container status \"f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901\": rpc error: code = NotFound desc = could not find container \"f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901\": container with ID starting with f758a22cf53eb4d4f1583230ab9ef69da745c7c570dce9e1500197da1ca92901 not found: ID does not exist" Dec 03 19:45:20 crc kubenswrapper[4758]: I1203 19:45:20.116490 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:45:20 crc kubenswrapper[4758]: E1203 19:45:20.116991 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:45:21 crc kubenswrapper[4758]: I1203 19:45:21.128204 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" path="/var/lib/kubelet/pods/cb1143fc-3616-410b-bb3e-1ba70ad27979/volumes" Dec 03 19:45:22 crc kubenswrapper[4758]: I1203 19:45:22.672027 4758 scope.go:117] "RemoveContainer" containerID="eed8eba36e8b27534e82423aabbe66e5ff3f5db85bc22fca0e643ccce51a59f0" Dec 03 19:45:32 crc kubenswrapper[4758]: I1203 19:45:32.118058 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:45:32 crc kubenswrapper[4758]: E1203 19:45:32.119016 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:45:43 crc kubenswrapper[4758]: I1203 19:45:43.115251 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:45:43 crc kubenswrapper[4758]: E1203 19:45:43.116236 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.778913 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l2gmm"] Dec 03 19:45:45 crc kubenswrapper[4758]: E1203 19:45:45.780052 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="extract-content" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.780071 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="extract-content" Dec 03 19:45:45 crc kubenswrapper[4758]: E1203 19:45:45.780118 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="extract-utilities" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.780127 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="extract-utilities" Dec 03 19:45:45 crc kubenswrapper[4758]: E1203 19:45:45.780141 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="registry-server" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.780147 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="registry-server" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.780387 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb1143fc-3616-410b-bb3e-1ba70ad27979" containerName="registry-server" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.782620 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.796461 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l2gmm"] Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.912438 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnk6z\" (UniqueName: \"kubernetes.io/projected/07350424-007f-4130-aec1-63081bacf607-kube-api-access-tnk6z\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.912745 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-utilities\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:45 crc kubenswrapper[4758]: I1203 19:45:45.912999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-catalog-content\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.015620 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-catalog-content\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.015940 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnk6z\" (UniqueName: \"kubernetes.io/projected/07350424-007f-4130-aec1-63081bacf607-kube-api-access-tnk6z\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.016008 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-utilities\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.016619 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-utilities\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.016866 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-catalog-content\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.044218 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnk6z\" (UniqueName: \"kubernetes.io/projected/07350424-007f-4130-aec1-63081bacf607-kube-api-access-tnk6z\") pod \"community-operators-l2gmm\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.115261 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:46 crc kubenswrapper[4758]: I1203 19:45:46.731752 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l2gmm"] Dec 03 19:45:47 crc kubenswrapper[4758]: I1203 19:45:47.222399 4758 generic.go:334] "Generic (PLEG): container finished" podID="07350424-007f-4130-aec1-63081bacf607" containerID="7533268c7914e3d4e95c435959a2e72fafdc81f3a97495662b0dddbf64c6f343" exitCode=0 Dec 03 19:45:47 crc kubenswrapper[4758]: I1203 19:45:47.222445 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerDied","Data":"7533268c7914e3d4e95c435959a2e72fafdc81f3a97495662b0dddbf64c6f343"} Dec 03 19:45:47 crc kubenswrapper[4758]: I1203 19:45:47.222646 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerStarted","Data":"e2dab2042315605373d480e45280d4ef041755113bdd98c360458830faaaccb8"} Dec 03 19:45:47 crc kubenswrapper[4758]: I1203 19:45:47.225068 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:45:50 crc kubenswrapper[4758]: I1203 19:45:50.260722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerStarted","Data":"304b9d583efd9c3352ca6cda9d44f60dc0c04d2373379ec8c2a59202978b947d"} Dec 03 19:45:52 crc kubenswrapper[4758]: I1203 19:45:52.285739 4758 generic.go:334] "Generic (PLEG): container finished" podID="07350424-007f-4130-aec1-63081bacf607" containerID="304b9d583efd9c3352ca6cda9d44f60dc0c04d2373379ec8c2a59202978b947d" exitCode=0 Dec 03 19:45:52 crc kubenswrapper[4758]: I1203 19:45:52.285797 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerDied","Data":"304b9d583efd9c3352ca6cda9d44f60dc0c04d2373379ec8c2a59202978b947d"} Dec 03 19:45:53 crc kubenswrapper[4758]: I1203 19:45:53.301803 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerStarted","Data":"2131539affa974ba3fea9baa97ba6f3f5ddeccc813eb711cce376bd461d71c24"} Dec 03 19:45:53 crc kubenswrapper[4758]: I1203 19:45:53.342254 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l2gmm" podStartSLOduration=2.860859075 podStartE2EDuration="8.342226481s" podCreationTimestamp="2025-12-03 19:45:45 +0000 UTC" firstStartedPulling="2025-12-03 19:45:47.224820252 +0000 UTC m=+10202.426197113" lastFinishedPulling="2025-12-03 19:45:52.706187658 +0000 UTC m=+10207.907564519" observedRunningTime="2025-12-03 19:45:53.329624176 +0000 UTC m=+10208.531001027" watchObservedRunningTime="2025-12-03 19:45:53.342226481 +0000 UTC m=+10208.543603342" Dec 03 19:45:56 crc kubenswrapper[4758]: I1203 19:45:56.115593 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:56 crc kubenswrapper[4758]: I1203 19:45:56.116164 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:56 crc kubenswrapper[4758]: I1203 19:45:56.174074 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:45:57 crc kubenswrapper[4758]: I1203 19:45:57.115056 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:45:57 crc kubenswrapper[4758]: E1203 19:45:57.115399 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:46:06 crc kubenswrapper[4758]: I1203 19:46:06.173879 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:46:06 crc kubenswrapper[4758]: I1203 19:46:06.241996 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l2gmm"] Dec 03 19:46:06 crc kubenswrapper[4758]: I1203 19:46:06.452784 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l2gmm" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="registry-server" containerID="cri-o://2131539affa974ba3fea9baa97ba6f3f5ddeccc813eb711cce376bd461d71c24" gracePeriod=2 Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.471224 4758 generic.go:334] "Generic (PLEG): container finished" podID="07350424-007f-4130-aec1-63081bacf607" containerID="2131539affa974ba3fea9baa97ba6f3f5ddeccc813eb711cce376bd461d71c24" exitCode=0 Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.471302 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerDied","Data":"2131539affa974ba3fea9baa97ba6f3f5ddeccc813eb711cce376bd461d71c24"} Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.662390 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.742745 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-utilities\") pod \"07350424-007f-4130-aec1-63081bacf607\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.742957 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-catalog-content\") pod \"07350424-007f-4130-aec1-63081bacf607\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.743048 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnk6z\" (UniqueName: \"kubernetes.io/projected/07350424-007f-4130-aec1-63081bacf607-kube-api-access-tnk6z\") pod \"07350424-007f-4130-aec1-63081bacf607\" (UID: \"07350424-007f-4130-aec1-63081bacf607\") " Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.743886 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-utilities" (OuterVolumeSpecName: "utilities") pod "07350424-007f-4130-aec1-63081bacf607" (UID: "07350424-007f-4130-aec1-63081bacf607"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.753004 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07350424-007f-4130-aec1-63081bacf607-kube-api-access-tnk6z" (OuterVolumeSpecName: "kube-api-access-tnk6z") pod "07350424-007f-4130-aec1-63081bacf607" (UID: "07350424-007f-4130-aec1-63081bacf607"). InnerVolumeSpecName "kube-api-access-tnk6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.808523 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07350424-007f-4130-aec1-63081bacf607" (UID: "07350424-007f-4130-aec1-63081bacf607"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.845977 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.846021 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07350424-007f-4130-aec1-63081bacf607-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:46:07 crc kubenswrapper[4758]: I1203 19:46:07.846036 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnk6z\" (UniqueName: \"kubernetes.io/projected/07350424-007f-4130-aec1-63081bacf607-kube-api-access-tnk6z\") on node \"crc\" DevicePath \"\"" Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.484242 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2gmm" event={"ID":"07350424-007f-4130-aec1-63081bacf607","Type":"ContainerDied","Data":"e2dab2042315605373d480e45280d4ef041755113bdd98c360458830faaaccb8"} Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.484312 4758 scope.go:117] "RemoveContainer" containerID="2131539affa974ba3fea9baa97ba6f3f5ddeccc813eb711cce376bd461d71c24" Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.484340 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2gmm" Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.519913 4758 scope.go:117] "RemoveContainer" containerID="304b9d583efd9c3352ca6cda9d44f60dc0c04d2373379ec8c2a59202978b947d" Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.535862 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l2gmm"] Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.557994 4758 scope.go:117] "RemoveContainer" containerID="7533268c7914e3d4e95c435959a2e72fafdc81f3a97495662b0dddbf64c6f343" Dec 03 19:46:08 crc kubenswrapper[4758]: I1203 19:46:08.564799 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l2gmm"] Dec 03 19:46:09 crc kubenswrapper[4758]: I1203 19:46:09.114921 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:46:09 crc kubenswrapper[4758]: E1203 19:46:09.115504 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:46:09 crc kubenswrapper[4758]: I1203 19:46:09.127629 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07350424-007f-4130-aec1-63081bacf607" path="/var/lib/kubelet/pods/07350424-007f-4130-aec1-63081bacf607/volumes" Dec 03 19:46:22 crc kubenswrapper[4758]: I1203 19:46:22.114944 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:46:22 crc kubenswrapper[4758]: E1203 19:46:22.115795 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:46:35 crc kubenswrapper[4758]: I1203 19:46:35.138081 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:46:35 crc kubenswrapper[4758]: E1203 19:46:35.138892 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.499112 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pcq4k"] Dec 03 19:46:47 crc kubenswrapper[4758]: E1203 19:46:47.500375 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="extract-utilities" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.500396 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="extract-utilities" Dec 03 19:46:47 crc kubenswrapper[4758]: E1203 19:46:47.500413 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="registry-server" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.500422 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="registry-server" Dec 03 19:46:47 crc kubenswrapper[4758]: E1203 19:46:47.500449 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="extract-content" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.500458 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="extract-content" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.500746 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="07350424-007f-4130-aec1-63081bacf607" containerName="registry-server" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.502870 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.519240 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcq4k"] Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.562048 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-catalog-content\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.562122 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjzpv\" (UniqueName: \"kubernetes.io/projected/ed061115-c734-48b2-b201-2e88f72b1ede-kube-api-access-xjzpv\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.562279 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-utilities\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.664710 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-catalog-content\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.664837 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjzpv\" (UniqueName: \"kubernetes.io/projected/ed061115-c734-48b2-b201-2e88f72b1ede-kube-api-access-xjzpv\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.664993 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-utilities\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.665649 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-utilities\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.666030 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-catalog-content\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.695283 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjzpv\" (UniqueName: \"kubernetes.io/projected/ed061115-c734-48b2-b201-2e88f72b1ede-kube-api-access-xjzpv\") pod \"redhat-marketplace-pcq4k\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:47 crc kubenswrapper[4758]: I1203 19:46:47.839244 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:48 crc kubenswrapper[4758]: I1203 19:46:48.114323 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:46:48 crc kubenswrapper[4758]: E1203 19:46:48.115120 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:46:48 crc kubenswrapper[4758]: I1203 19:46:48.363723 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcq4k"] Dec 03 19:46:48 crc kubenswrapper[4758]: I1203 19:46:48.912360 4758 generic.go:334] "Generic (PLEG): container finished" podID="ed061115-c734-48b2-b201-2e88f72b1ede" containerID="cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807" exitCode=0 Dec 03 19:46:48 crc kubenswrapper[4758]: I1203 19:46:48.912418 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcq4k" event={"ID":"ed061115-c734-48b2-b201-2e88f72b1ede","Type":"ContainerDied","Data":"cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807"} Dec 03 19:46:48 crc kubenswrapper[4758]: I1203 19:46:48.912484 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcq4k" event={"ID":"ed061115-c734-48b2-b201-2e88f72b1ede","Type":"ContainerStarted","Data":"5a0fb08511787a59f055680e33a5c17a921fe9adf123c495623d58f83c98e218"} Dec 03 19:46:50 crc kubenswrapper[4758]: I1203 19:46:50.937335 4758 generic.go:334] "Generic (PLEG): container finished" podID="ed061115-c734-48b2-b201-2e88f72b1ede" containerID="a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13" exitCode=0 Dec 03 19:46:50 crc kubenswrapper[4758]: I1203 19:46:50.937395 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcq4k" event={"ID":"ed061115-c734-48b2-b201-2e88f72b1ede","Type":"ContainerDied","Data":"a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13"} Dec 03 19:46:51 crc kubenswrapper[4758]: I1203 19:46:51.951747 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcq4k" event={"ID":"ed061115-c734-48b2-b201-2e88f72b1ede","Type":"ContainerStarted","Data":"7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97"} Dec 03 19:46:51 crc kubenswrapper[4758]: I1203 19:46:51.985578 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pcq4k" podStartSLOduration=2.411300391 podStartE2EDuration="4.985555337s" podCreationTimestamp="2025-12-03 19:46:47 +0000 UTC" firstStartedPulling="2025-12-03 19:46:48.916094956 +0000 UTC m=+10264.117471817" lastFinishedPulling="2025-12-03 19:46:51.490349902 +0000 UTC m=+10266.691726763" observedRunningTime="2025-12-03 19:46:51.974522733 +0000 UTC m=+10267.175899604" watchObservedRunningTime="2025-12-03 19:46:51.985555337 +0000 UTC m=+10267.186932198" Dec 03 19:46:57 crc kubenswrapper[4758]: I1203 19:46:57.839888 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:57 crc kubenswrapper[4758]: I1203 19:46:57.840549 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:57 crc kubenswrapper[4758]: I1203 19:46:57.894138 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:58 crc kubenswrapper[4758]: I1203 19:46:58.063540 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:46:58 crc kubenswrapper[4758]: I1203 19:46:58.134152 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcq4k"] Dec 03 19:47:00 crc kubenswrapper[4758]: I1203 19:47:00.029982 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pcq4k" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="registry-server" containerID="cri-o://7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97" gracePeriod=2 Dec 03 19:47:00 crc kubenswrapper[4758]: I1203 19:47:00.856786 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:47:00 crc kubenswrapper[4758]: I1203 19:47:00.994429 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjzpv\" (UniqueName: \"kubernetes.io/projected/ed061115-c734-48b2-b201-2e88f72b1ede-kube-api-access-xjzpv\") pod \"ed061115-c734-48b2-b201-2e88f72b1ede\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " Dec 03 19:47:00 crc kubenswrapper[4758]: I1203 19:47:00.994530 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-catalog-content\") pod \"ed061115-c734-48b2-b201-2e88f72b1ede\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " Dec 03 19:47:00 crc kubenswrapper[4758]: I1203 19:47:00.994797 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-utilities\") pod \"ed061115-c734-48b2-b201-2e88f72b1ede\" (UID: \"ed061115-c734-48b2-b201-2e88f72b1ede\") " Dec 03 19:47:00 crc kubenswrapper[4758]: I1203 19:47:00.995755 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-utilities" (OuterVolumeSpecName: "utilities") pod "ed061115-c734-48b2-b201-2e88f72b1ede" (UID: "ed061115-c734-48b2-b201-2e88f72b1ede"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.000927 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed061115-c734-48b2-b201-2e88f72b1ede-kube-api-access-xjzpv" (OuterVolumeSpecName: "kube-api-access-xjzpv") pod "ed061115-c734-48b2-b201-2e88f72b1ede" (UID: "ed061115-c734-48b2-b201-2e88f72b1ede"). InnerVolumeSpecName "kube-api-access-xjzpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.018273 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed061115-c734-48b2-b201-2e88f72b1ede" (UID: "ed061115-c734-48b2-b201-2e88f72b1ede"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.043386 4758 generic.go:334] "Generic (PLEG): container finished" podID="ed061115-c734-48b2-b201-2e88f72b1ede" containerID="7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97" exitCode=0 Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.043435 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcq4k" event={"ID":"ed061115-c734-48b2-b201-2e88f72b1ede","Type":"ContainerDied","Data":"7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97"} Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.043464 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcq4k" event={"ID":"ed061115-c734-48b2-b201-2e88f72b1ede","Type":"ContainerDied","Data":"5a0fb08511787a59f055680e33a5c17a921fe9adf123c495623d58f83c98e218"} Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.043484 4758 scope.go:117] "RemoveContainer" containerID="7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.043645 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcq4k" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.097548 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.097925 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjzpv\" (UniqueName: \"kubernetes.io/projected/ed061115-c734-48b2-b201-2e88f72b1ede-kube-api-access-xjzpv\") on node \"crc\" DevicePath \"\"" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.097937 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed061115-c734-48b2-b201-2e88f72b1ede-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.099399 4758 scope.go:117] "RemoveContainer" containerID="a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.099874 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcq4k"] Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.111577 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcq4k"] Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.114515 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:47:01 crc kubenswrapper[4758]: E1203 19:47:01.114993 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.127399 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" path="/var/lib/kubelet/pods/ed061115-c734-48b2-b201-2e88f72b1ede/volumes" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.128387 4758 scope.go:117] "RemoveContainer" containerID="cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.176073 4758 scope.go:117] "RemoveContainer" containerID="7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97" Dec 03 19:47:01 crc kubenswrapper[4758]: E1203 19:47:01.177466 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97\": container with ID starting with 7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97 not found: ID does not exist" containerID="7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.177557 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97"} err="failed to get container status \"7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97\": rpc error: code = NotFound desc = could not find container \"7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97\": container with ID starting with 7ab8b9366a412b3ca6ff8912d5e4c6e2e1081a6ae3a4b9083cb4e91e7f1a1f97 not found: ID does not exist" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.177600 4758 scope.go:117] "RemoveContainer" containerID="a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13" Dec 03 19:47:01 crc kubenswrapper[4758]: E1203 19:47:01.178259 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13\": container with ID starting with a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13 not found: ID does not exist" containerID="a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.178357 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13"} err="failed to get container status \"a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13\": rpc error: code = NotFound desc = could not find container \"a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13\": container with ID starting with a842806e97a53794908d27ab02ddbbe81b430ade866efa35077b6b4002bb2a13 not found: ID does not exist" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.178392 4758 scope.go:117] "RemoveContainer" containerID="cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807" Dec 03 19:47:01 crc kubenswrapper[4758]: E1203 19:47:01.178975 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807\": container with ID starting with cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807 not found: ID does not exist" containerID="cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807" Dec 03 19:47:01 crc kubenswrapper[4758]: I1203 19:47:01.179037 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807"} err="failed to get container status \"cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807\": rpc error: code = NotFound desc = could not find container \"cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807\": container with ID starting with cbeb2fcce06196159812747a27b67b5be71daf3c63d1f833a86f5a25ad7e3807 not found: ID does not exist" Dec 03 19:47:14 crc kubenswrapper[4758]: I1203 19:47:14.115132 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:47:14 crc kubenswrapper[4758]: E1203 19:47:14.117183 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:47:22 crc kubenswrapper[4758]: I1203 19:47:22.799811 4758 scope.go:117] "RemoveContainer" containerID="7f4f463f05ef45138fdd84da78fd57eb36efcffd72d20498b7a09d206a75a2c5" Dec 03 19:47:27 crc kubenswrapper[4758]: I1203 19:47:27.115141 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:47:27 crc kubenswrapper[4758]: E1203 19:47:27.116041 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.054300 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0850e686-52c2-42fd-bd78-7c0feef9ac6d/init-config-reloader/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.312148 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0850e686-52c2-42fd-bd78-7c0feef9ac6d/alertmanager/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.366129 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0850e686-52c2-42fd-bd78-7c0feef9ac6d/init-config-reloader/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.372605 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0850e686-52c2-42fd-bd78-7c0feef9ac6d/config-reloader/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.626059 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ccfcd869-afcd-4a92-a02c-83b861bf54f8/aodh-evaluator/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.686182 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ccfcd869-afcd-4a92-a02c-83b861bf54f8/aodh-api/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.734357 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ccfcd869-afcd-4a92-a02c-83b861bf54f8/aodh-listener/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.857988 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ccfcd869-afcd-4a92-a02c-83b861bf54f8/aodh-notifier/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.948472 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5f468494d-9f2zm_947fb70a-1772-4b0e-b6b2-eff63219feeb/barbican-api/0.log" Dec 03 19:47:38 crc kubenswrapper[4758]: I1203 19:47:38.999708 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5f468494d-9f2zm_947fb70a-1772-4b0e-b6b2-eff63219feeb/barbican-api-log/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.259136 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b5dbc78b-xh7jg_1f3421fd-2c4e-4ec0-9bca-345973cc7fb0/barbican-keystone-listener-log/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.298591 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b5dbc78b-xh7jg_1f3421fd-2c4e-4ec0-9bca-345973cc7fb0/barbican-keystone-listener/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.447650 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7f48c6d99f-5k4sg_b740aeeb-6458-4a97-a7bc-a486f0ed2d97/barbican-worker/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.468069 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7f48c6d99f-5k4sg_b740aeeb-6458-4a97-a7bc-a486f0ed2d97/barbican-worker-log/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.607244 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-jwtcm_cbe9768c-6273-4788-8aa7-8ed927817cc5/bootstrap-openstack-openstack-cell1/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.798992 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f81e7681-d858-4cf2-97dc-08a6e8c0b6f6/ceilometer-central-agent/1.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.851887 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f81e7681-d858-4cf2-97dc-08a6e8c0b6f6/ceilometer-central-agent/0.log" Dec 03 19:47:39 crc kubenswrapper[4758]: I1203 19:47:39.891996 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f81e7681-d858-4cf2-97dc-08a6e8c0b6f6/ceilometer-notification-agent/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.008322 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f81e7681-d858-4cf2-97dc-08a6e8c0b6f6/proxy-httpd/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.034538 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f81e7681-d858-4cf2-97dc-08a6e8c0b6f6/sg-core/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.122903 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-649l7_bf9c857e-ddc6-41fb-bc6c-9fe40977bc92/ceph-client-openstack-openstack-cell1/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.335181 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10679416-8851-4d9c-9450-6722e89c06ba/cinder-api/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.362649 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10679416-8851-4d9c-9450-6722e89c06ba/cinder-api-log/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.644804 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_e783b83e-d408-4f5b-b2ba-c322f8bd93f5/probe/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.661726 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_e783b83e-d408-4f5b-b2ba-c322f8bd93f5/cinder-backup/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.695602 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8d01ab7c-bf47-4378-b41c-508c5e71e0b4/cinder-scheduler/0.log" Dec 03 19:47:40 crc kubenswrapper[4758]: I1203 19:47:40.896546 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8d01ab7c-bf47-4378-b41c-508c5e71e0b4/probe/0.log" Dec 03 19:47:41 crc kubenswrapper[4758]: I1203 19:47:41.006084 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_779faeb6-b6e1-4953-9c93-41fc7e90dcb7/cinder-volume/0.log" Dec 03 19:47:41 crc kubenswrapper[4758]: I1203 19:47:41.065669 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_779faeb6-b6e1-4953-9c93-41fc7e90dcb7/probe/0.log" Dec 03 19:47:41 crc kubenswrapper[4758]: I1203 19:47:41.257134 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-82l6j_5d46d38d-5c92-409c-bc14-578aca300889/configure-network-openstack-openstack-cell1/0.log" Dec 03 19:47:41 crc kubenswrapper[4758]: I1203 19:47:41.321001 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-nlm2b_aa20be19-9e47-498e-b64f-a4840514eaae/configure-os-openstack-openstack-cell1/0.log" Dec 03 19:47:41 crc kubenswrapper[4758]: I1203 19:47:41.529973 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8444f75b9f-zf8xz_68c8bfdf-3a84-4c3b-8a8d-1df245208200/init/0.log" Dec 03 19:47:41 crc kubenswrapper[4758]: I1203 19:47:41.938051 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8444f75b9f-zf8xz_68c8bfdf-3a84-4c3b-8a8d-1df245208200/init/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.115190 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:47:42 crc kubenswrapper[4758]: E1203 19:47:42.115552 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.136267 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-vlgwn_8f4a384c-e511-43ab-b0e0-8def30abdcd4/download-cache-openstack-openstack-cell1/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.156198 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8444f75b9f-zf8xz_68c8bfdf-3a84-4c3b-8a8d-1df245208200/dnsmasq-dns/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.288642 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aca4a33e-8dac-4d3b-bd2b-7ef196a94871/glance-httpd/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.387937 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aca4a33e-8dac-4d3b-bd2b-7ef196a94871/glance-log/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.546997 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_579d0799-97e3-468a-afcb-6c0ca7c67004/glance-httpd/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.560817 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_579d0799-97e3-468a-afcb-6c0ca7c67004/glance-log/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.813782 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-cd9dd694-vjtvk_bede3802-5d16-4df3-8595-4af3c5be0272/heat-api/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.924168 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6f84647496-588vd_9121c214-9012-49c7-8f48-34f2327d3cb9/heat-engine/0.log" Dec 03 19:47:42 crc kubenswrapper[4758]: I1203 19:47:42.932790 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-7bf88cbcf7-7tk6c_508aa60b-f57e-46f2-97f4-43e7c7bedf31/heat-cfnapi/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.127970 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6654448665-7d6rf_223e253a-8088-4569-935b-7f7bed328067/horizon/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.156489 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-wp698_29991f51-5ebf-4690-9196-7bb5a47202e9/install-certs-openstack-openstack-cell1/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.255249 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6654448665-7d6rf_223e253a-8088-4569-935b-7f7bed328067/horizon-log/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.406161 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-88rh2_ca4d63dd-6c41-43c0-9b30-fcece1c986b5/install-os-openstack-openstack-cell1/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.568202 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413141-k86xb_1ee10a9b-0fcf-4701-802d-480e6131feed/keystone-cron/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.597285 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7669c5b48c-b2qhl_7257c7da-68c8-4cbf-b7e4-c332ebc2ff2f/keystone-api/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.722146 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_07dd279b-b4fa-4544-9afe-1e0639826a54/kube-state-metrics/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.820959 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-wkqhj_0594107d-92c8-4f63-9d64-5c362703c627/libvirt-openstack-openstack-cell1/0.log" Dec 03 19:47:43 crc kubenswrapper[4758]: I1203 19:47:43.972412 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_068d058a-40e9-4d0c-bd12-56290e862552/manila-api-log/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.050646 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_068d058a-40e9-4d0c-bd12-56290e862552/manila-api/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.145023 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d/probe/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.202516 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_1ebcd8dd-3ef4-4cb7-90fd-0db58cc1dd8d/manila-scheduler/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.301771 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e7f8de8f-60ad-4b7e-92c5-6df43aae3589/manila-share/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.357714 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e7f8de8f-60ad-4b7e-92c5-6df43aae3589/probe/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.681642 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d7df67465-q66s8_4870a274-a501-43ce-89f6-6d974d05ce8c/neutron-httpd/0.log" Dec 03 19:47:44 crc kubenswrapper[4758]: I1203 19:47:44.728520 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d7df67465-q66s8_4870a274-a501-43ce-89f6-6d974d05ce8c/neutron-api/0.log" Dec 03 19:47:45 crc kubenswrapper[4758]: I1203 19:47:45.394530 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-hpqxx_4689d0b4-d192-4b64-bcb7-937098f1f2ee/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 03 19:47:45 crc kubenswrapper[4758]: I1203 19:47:45.543042 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-xrh25_1eb85ae0-647a-4e35-a8bb-404f07ee1eee/neutron-metadata-openstack-openstack-cell1/0.log" Dec 03 19:47:45 crc kubenswrapper[4758]: I1203 19:47:45.774245 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-tzxnl_80659ad8-2251-46bb-853d-18ac040f6d84/neutron-sriov-openstack-openstack-cell1/0.log" Dec 03 19:47:45 crc kubenswrapper[4758]: I1203 19:47:45.901112 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c5e024a3-38fa-4d28-9276-b4cf52206784/nova-api-api/0.log" Dec 03 19:47:46 crc kubenswrapper[4758]: I1203 19:47:46.067069 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c5e024a3-38fa-4d28-9276-b4cf52206784/nova-api-log/0.log" Dec 03 19:47:46 crc kubenswrapper[4758]: I1203 19:47:46.211922 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_56369ab2-3a2d-4164-8801-de6ad870b6fa/nova-cell0-conductor-conductor/0.log" Dec 03 19:47:46 crc kubenswrapper[4758]: I1203 19:47:46.443434 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e8bc8eb0-3634-4e45-9d75-5023116c4585/nova-cell1-conductor-conductor/0.log" Dec 03 19:47:46 crc kubenswrapper[4758]: I1203 19:47:46.577217 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_dfc3714e-e415-4d7a-8710-9050fdae8e0b/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 19:47:46 crc kubenswrapper[4758]: I1203 19:47:46.779780 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellvn52v_f5a21acf-8141-463b-855f-959e09815eee/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 03 19:47:46 crc kubenswrapper[4758]: I1203 19:47:46.892375 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-bsb6m_d6d965b5-a0a8-4ba8-a25c-623612615e02/nova-cell1-openstack-openstack-cell1/0.log" Dec 03 19:47:47 crc kubenswrapper[4758]: I1203 19:47:47.935929 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_01a675e1-3969-4c03-8026-5c7206ed8dee/nova-metadata-metadata/0.log" Dec 03 19:47:47 crc kubenswrapper[4758]: I1203 19:47:47.991850 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_01a675e1-3969-4c03-8026-5c7206ed8dee/nova-metadata-log/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.010529 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d3860513-f195-43f5-af71-26cb0e5a37a8/nova-scheduler-scheduler/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.262075 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-78cc97546b-97hmr_9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1/init/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.487643 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-78cc97546b-97hmr_9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1/init/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.531937 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-78cc97546b-97hmr_9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1/octavia-api-provider-agent/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.753820 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-8wngx_8d1e1443-d203-4e87-8a80-f2d9785eb981/init/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.776567 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-78cc97546b-97hmr_9f6d1d6d-811b-4bc4-a0f4-95424a9c6bd1/octavia-api/0.log" Dec 03 19:47:48 crc kubenswrapper[4758]: I1203 19:47:48.993851 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-8wngx_8d1e1443-d203-4e87-8a80-f2d9785eb981/init/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.112326 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-tdpvl_1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907/init/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.160549 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-8wngx_8d1e1443-d203-4e87-8a80-f2d9785eb981/octavia-healthmanager/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.371587 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-tdpvl_1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907/init/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.377932 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-tdpvl_1a2a8136-2ce0-4cb9-9bae-b2ea4dbd3907/octavia-housekeeping/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.488037 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-mfngn_397e1e84-cff5-44c1-9ce9-1ba6022b95db/init/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.736552 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-mfngn_397e1e84-cff5-44c1-9ce9-1ba6022b95db/octavia-amphora-httpd/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.752925 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-mfngn_397e1e84-cff5-44c1-9ce9-1ba6022b95db/init/0.log" Dec 03 19:47:49 crc kubenswrapper[4758]: I1203 19:47:49.788848 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-9d5h7_e0be9d87-ab3b-49f5-b103-7aa960127e4b/init/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.092533 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-9d5h7_e0be9d87-ab3b-49f5-b103-7aa960127e4b/init/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.157816 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-l9dgm_e4c53e2b-203e-4767-9943-d063ef13ff7c/init/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.180475 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-9d5h7_e0be9d87-ab3b-49f5-b103-7aa960127e4b/octavia-rsyslog/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.355893 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-l9dgm_e4c53e2b-203e-4767-9943-d063ef13ff7c/init/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.500574 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0bb80e50-176a-4fc0-890d-026294efeaaa/mysql-bootstrap/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.598582 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-l9dgm_e4c53e2b-203e-4767-9943-d063ef13ff7c/octavia-worker/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.686092 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0bb80e50-176a-4fc0-890d-026294efeaaa/mysql-bootstrap/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.804171 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0bb80e50-176a-4fc0-890d-026294efeaaa/galera/0.log" Dec 03 19:47:50 crc kubenswrapper[4758]: I1203 19:47:50.936498 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3/mysql-bootstrap/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.247927 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3/mysql-bootstrap/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.267215 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_bac9f3c0-3dc6-4edd-ab4e-3d45e17bcc75/openstackclient/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.303214 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_cdd16c7d-f0b9-4d78-afcf-8f1ce16acec3/galera/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.507614 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hzhjn_a86f6acf-8cd6-41a9-a38a-1f113fe29d93/openstack-network-exporter/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.616861 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5jrck_b829c7e2-ce7c-46f9-830b-aa7227537d2c/ovsdb-server-init/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.813237 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5jrck_b829c7e2-ce7c-46f9-830b-aa7227537d2c/ovsdb-server-init/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.937396 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5jrck_b829c7e2-ce7c-46f9-830b-aa7227537d2c/ovs-vswitchd/0.log" Dec 03 19:47:51 crc kubenswrapper[4758]: I1203 19:47:51.969188 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5jrck_b829c7e2-ce7c-46f9-830b-aa7227537d2c/ovsdb-server/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.063168 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vts8k_45509ece-d0bc-4fed-aaea-825badb17447/ovn-controller/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.263271 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_41e72dc3-c313-42dd-a31e-c7bbcf63ff5a/ovn-northd/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.273997 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_41e72dc3-c313-42dd-a31e-c7bbcf63ff5a/openstack-network-exporter/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.532390 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_de936fcf-6748-44f0-ad21-e62bec06dba8/openstack-network-exporter/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.584460 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-6mrsp_92c8b34e-4a7c-49ad-8678-b9d85206ee55/ovn-openstack-openstack-cell1/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.659314 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_de936fcf-6748-44f0-ad21-e62bec06dba8/ovsdbserver-nb/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.782527 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_49778bc2-4f74-4f86-9558-f15d22013db1/openstack-network-exporter/0.log" Dec 03 19:47:52 crc kubenswrapper[4758]: I1203 19:47:52.864359 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_49778bc2-4f74-4f86-9558-f15d22013db1/ovsdbserver-nb/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.013568 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_51e5a5cf-4500-43de-90bf-788460adcfe5/openstack-network-exporter/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.083157 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_51e5a5cf-4500-43de-90bf-788460adcfe5/ovsdbserver-nb/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.114849 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:47:53 crc kubenswrapper[4758]: E1203 19:47:53.115227 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.190296 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6812355e-c51f-499b-bfe0-3d78d40e86c7/openstack-network-exporter/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.277299 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6812355e-c51f-499b-bfe0-3d78d40e86c7/ovsdbserver-sb/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.428580 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_a7d6fb8c-07d0-41f4-b027-76d126425a78/ovsdbserver-sb/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.445477 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_a7d6fb8c-07d0-41f4-b027-76d126425a78/openstack-network-exporter/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.816032 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_3ba68c62-f0d6-4472-a7db-9c24b70fbb58/openstack-network-exporter/0.log" Dec 03 19:47:53 crc kubenswrapper[4758]: I1203 19:47:53.953367 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_3ba68c62-f0d6-4472-a7db-9c24b70fbb58/ovsdbserver-sb/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.251901 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-645b59b84b-kb6kd_99571942-c27b-4b35-8d59-3ba4450b2f3a/placement-api/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.271831 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-645b59b84b-kb6kd_99571942-c27b-4b35-8d59-3ba4450b2f3a/placement-log/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.355938 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cdfwks_92bcc00c-1a64-4bb1-8152-25f30ff242a9/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.522326 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_475fe5d8-4473-4c4f-a6ca-4f8a60a10060/init-config-reloader/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.731764 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_475fe5d8-4473-4c4f-a6ca-4f8a60a10060/config-reloader/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.809267 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_475fe5d8-4473-4c4f-a6ca-4f8a60a10060/init-config-reloader/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.837874 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_475fe5d8-4473-4c4f-a6ca-4f8a60a10060/prometheus/0.log" Dec 03 19:47:54 crc kubenswrapper[4758]: I1203 19:47:54.908664 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_475fe5d8-4473-4c4f-a6ca-4f8a60a10060/thanos-sidecar/0.log" Dec 03 19:47:55 crc kubenswrapper[4758]: I1203 19:47:55.065894 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6707c0c8-b302-46bf-a6d7-e953f5f2d497/setup-container/0.log" Dec 03 19:47:55 crc kubenswrapper[4758]: I1203 19:47:55.381767 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6707c0c8-b302-46bf-a6d7-e953f5f2d497/setup-container/0.log" Dec 03 19:47:55 crc kubenswrapper[4758]: I1203 19:47:55.427947 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a99f005-65b5-4d3b-a49f-75213f40b232/setup-container/0.log" Dec 03 19:47:55 crc kubenswrapper[4758]: I1203 19:47:55.489764 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6707c0c8-b302-46bf-a6d7-e953f5f2d497/rabbitmq/0.log" Dec 03 19:47:55 crc kubenswrapper[4758]: I1203 19:47:55.765630 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a99f005-65b5-4d3b-a49f-75213f40b232/setup-container/0.log" Dec 03 19:47:55 crc kubenswrapper[4758]: I1203 19:47:55.897999 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-k44x6_9934ab2d-58c4-44e7-8021-b74ec84341f9/reboot-os-openstack-openstack-cell1/0.log" Dec 03 19:47:56 crc kubenswrapper[4758]: I1203 19:47:56.152124 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-q9w8s_131aa418-2634-4ccf-8ae2-5ac72ecfc8eb/run-os-openstack-openstack-cell1/0.log" Dec 03 19:47:56 crc kubenswrapper[4758]: I1203 19:47:56.941876 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-ldt7n_cb99c7dc-9a7d-4287-b0f9-9590049db61d/ssh-known-hosts-openstack/0.log" Dec 03 19:47:57 crc kubenswrapper[4758]: I1203 19:47:57.059237 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-pbnq4_fa3c885d-c09e-4fdf-ae60-61bbfc92df32/telemetry-openstack-openstack-cell1/0.log" Dec 03 19:47:57 crc kubenswrapper[4758]: I1203 19:47:57.347717 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-c7q2m_650575e1-2ffe-41e4-887a-ae4e451bc53a/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 03 19:47:57 crc kubenswrapper[4758]: I1203 19:47:57.521574 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-x6gff_57d874e8-5585-46a0-973c-3cfa587a7e3d/validate-network-openstack-openstack-cell1/0.log" Dec 03 19:47:57 crc kubenswrapper[4758]: I1203 19:47:57.539388 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_141e0710-b827-4f55-af39-16ef9ffc7efd/memcached/0.log" Dec 03 19:47:59 crc kubenswrapper[4758]: I1203 19:47:59.045258 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4a99f005-65b5-4d3b-a49f-75213f40b232/rabbitmq/0.log" Dec 03 19:48:07 crc kubenswrapper[4758]: I1203 19:48:07.115250 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:48:07 crc kubenswrapper[4758]: E1203 19:48:07.116185 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:48:21 crc kubenswrapper[4758]: I1203 19:48:21.252846 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/util/0.log" Dec 03 19:48:21 crc kubenswrapper[4758]: I1203 19:48:21.399080 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/util/0.log" Dec 03 19:48:21 crc kubenswrapper[4758]: I1203 19:48:21.410226 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/pull/0.log" Dec 03 19:48:21 crc kubenswrapper[4758]: I1203 19:48:21.436816 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/pull/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.115543 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.139118 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/pull/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.147538 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/util/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.179957 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_432be7e02dc2ed7a306c43f99633d03549b7f319970fe11bd395585ef4q2vt4_43b518b0-93c4-45bc-8b3c-a997dd2a02ee/extract/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.385209 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-fnlpm_de31e4e2-fe86-4543-82c5-da042e77b6ce/kube-rbac-proxy/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.494359 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-fnlpm_de31e4e2-fe86-4543-82c5-da042e77b6ce/manager/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.539910 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j4sfr_36504cff-c657-485e-9db7-788382c4a9e1/kube-rbac-proxy/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.709496 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j4sfr_36504cff-c657-485e-9db7-788382c4a9e1/manager/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.750314 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-8ll9j_99462bc7-4e24-40b1-8ce5-555c47c8305f/kube-rbac-proxy/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.795831 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-8ll9j_99462bc7-4e24-40b1-8ce5-555c47c8305f/manager/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.952306 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-72tmg_84cea5a5-c77f-4818-9c41-8d91a838f7fa/kube-rbac-proxy/0.log" Dec 03 19:48:22 crc kubenswrapper[4758]: I1203 19:48:22.978971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"e782b9e9687f9787128adb5b32b6b6158f798ff4567b2a8122c1b1aad4fc6275"} Dec 03 19:48:23 crc kubenswrapper[4758]: I1203 19:48:23.165303 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-72tmg_84cea5a5-c77f-4818-9c41-8d91a838f7fa/manager/0.log" Dec 03 19:48:23 crc kubenswrapper[4758]: I1203 19:48:23.239207 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4972h_5477637e-ea68-454e-8cf6-51b491464daf/kube-rbac-proxy/0.log" Dec 03 19:48:23 crc kubenswrapper[4758]: I1203 19:48:23.304237 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4972h_5477637e-ea68-454e-8cf6-51b491464daf/manager/0.log" Dec 03 19:48:23 crc kubenswrapper[4758]: I1203 19:48:23.384571 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-q8ddt_61019cc3-111f-4f59-8e43-64b5dfdd22c4/kube-rbac-proxy/0.log" Dec 03 19:48:23 crc kubenswrapper[4758]: I1203 19:48:23.534030 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-q8ddt_61019cc3-111f-4f59-8e43-64b5dfdd22c4/manager/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.245882 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-7m28f_1911d89e-c35c-4104-b0c1-265f770a731d/kube-rbac-proxy/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.269473 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-5djcl_144cd5a6-c5c5-456d-b28d-27c3a3fb8898/kube-rbac-proxy/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.576765 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-5djcl_144cd5a6-c5c5-456d-b28d-27c3a3fb8898/manager/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.593885 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-r5r8j_4ecb6824-b536-4432-8e37-92dc55964815/kube-rbac-proxy/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.687248 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-7m28f_1911d89e-c35c-4104-b0c1-265f770a731d/manager/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.845008 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qcpt6_f6893abc-6551-4bc5-bcc5-18dd74da025b/kube-rbac-proxy/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.935994 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-r5r8j_4ecb6824-b536-4432-8e37-92dc55964815/manager/0.log" Dec 03 19:48:24 crc kubenswrapper[4758]: I1203 19:48:24.971277 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qcpt6_f6893abc-6551-4bc5-bcc5-18dd74da025b/manager/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.130497 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-ndmwx_74b430d9-3316-4157-ab4f-264f7797c7e9/kube-rbac-proxy/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.242406 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-ndmwx_74b430d9-3316-4157-ab4f-264f7797c7e9/manager/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.290145 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5mgxc_c624801b-66ee-4a15-866a-a65e7c69b2df/kube-rbac-proxy/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.390894 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-5mgxc_c624801b-66ee-4a15-866a-a65e7c69b2df/manager/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.460825 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-c5wfm_29c8457a-abba-43c0-b2ac-12c78bdeef4c/kube-rbac-proxy/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.616628 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-rp8pj_7806e6cc-eee5-4735-bae7-10ca5e4144d6/kube-rbac-proxy/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.764211 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-c5wfm_29c8457a-abba-43c0-b2ac-12c78bdeef4c/manager/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.764910 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-rp8pj_7806e6cc-eee5-4735-bae7-10ca5e4144d6/manager/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.852110 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j_ff85f702-0ab2-490b-899b-e14bcd599eb4/manager/0.log" Dec 03 19:48:25 crc kubenswrapper[4758]: I1203 19:48:25.859895 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4qh82j_ff85f702-0ab2-490b-899b-e14bcd599eb4/kube-rbac-proxy/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.208925 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6b5fc879f-9f9cq_b2f53a34-0ea1-4d23-8b26-a077003f0f9c/operator/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.403972 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-48xhl_ee9c8f83-7b21-4742-9e08-90c355c3c0fe/registry-server/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.450908 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-p8mlv_ce961712-a4d4-4a93-96e6-0cf4fdc58eea/kube-rbac-proxy/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.629301 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-p8mlv_ce961712-a4d4-4a93-96e6-0cf4fdc58eea/manager/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.654148 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4pfjv_aae66ae2-507f-43f8-bac5-c842427bb780/kube-rbac-proxy/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.789338 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4pfjv_aae66ae2-507f-43f8-bac5-c842427bb780/manager/0.log" Dec 03 19:48:26 crc kubenswrapper[4758]: I1203 19:48:26.977377 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-4vmjv_5a38f83d-d017-49fc-98bf-e08ae999428c/operator/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.075966 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-2j99q_eefa346f-ce5f-4445-8565-f6ad8250f049/kube-rbac-proxy/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.172551 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-2j99q_eefa346f-ce5f-4445-8565-f6ad8250f049/manager/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.263742 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-5x7rp_61d9456a-70b6-4a3c-80fe-1ad9a65d0f44/kube-rbac-proxy/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.502821 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-54w5p_56b5e5d0-a007-4451-bfbb-87e6f7201367/kube-rbac-proxy/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.602186 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-54w5p_56b5e5d0-a007-4451-bfbb-87e6f7201367/manager/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.622752 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-5x7rp_61d9456a-70b6-4a3c-80fe-1ad9a65d0f44/manager/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.895289 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-6llwb_68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1/kube-rbac-proxy/0.log" Dec 03 19:48:27 crc kubenswrapper[4758]: I1203 19:48:27.922448 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-6llwb_68b7f7c2-bdf0-44e9-aee9-5e2e700a6ca1/manager/0.log" Dec 03 19:48:28 crc kubenswrapper[4758]: I1203 19:48:28.790309 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-84685d4b7d-4f4l6_ee20d8c6-5b9b-4918-ac2c-4f30ca94c37c/manager/0.log" Dec 03 19:48:52 crc kubenswrapper[4758]: I1203 19:48:52.222159 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-nx6lj_fbb99ef5-6a95-43f4-a404-1d4a14cf28e8/control-plane-machine-set-operator/0.log" Dec 03 19:48:52 crc kubenswrapper[4758]: I1203 19:48:52.340539 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7lnzm_014604aa-7ce2-48ed-b924-00ac8db7abb1/kube-rbac-proxy/0.log" Dec 03 19:48:52 crc kubenswrapper[4758]: I1203 19:48:52.431626 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7lnzm_014604aa-7ce2-48ed-b924-00ac8db7abb1/machine-api-operator/0.log" Dec 03 19:49:07 crc kubenswrapper[4758]: I1203 19:49:07.403124 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-q85ck_8cb254c1-10c6-4f10-b201-85f3b5b66b03/cert-manager-controller/0.log" Dec 03 19:49:07 crc kubenswrapper[4758]: I1203 19:49:07.505348 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-wlvn2_cefd882c-e7fc-4a9e-bcb4-2d349b184f7c/cert-manager-cainjector/0.log" Dec 03 19:49:07 crc kubenswrapper[4758]: I1203 19:49:07.598946 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-rf2cm_bf8ec7f2-ab50-4adc-b3f2-f19511690ddd/cert-manager-webhook/0.log" Dec 03 19:49:22 crc kubenswrapper[4758]: I1203 19:49:22.962316 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-cznmb_4eff08a5-e932-4488-9f0f-4a4584ec8c11/nmstate-console-plugin/0.log" Dec 03 19:49:23 crc kubenswrapper[4758]: I1203 19:49:23.218966 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-xfg7s_dd83999a-a4d2-4880-ab08-2e73e2eddd21/nmstate-handler/0.log" Dec 03 19:49:23 crc kubenswrapper[4758]: I1203 19:49:23.231857 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-c92pg_edb53861-39ab-4c5c-ac89-7cd726ce6dcc/kube-rbac-proxy/0.log" Dec 03 19:49:23 crc kubenswrapper[4758]: I1203 19:49:23.311583 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-c92pg_edb53861-39ab-4c5c-ac89-7cd726ce6dcc/nmstate-metrics/0.log" Dec 03 19:49:23 crc kubenswrapper[4758]: I1203 19:49:23.909249 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-tjzhr_7267aad3-2922-44b5-85b2-4c40d43c5849/nmstate-webhook/0.log" Dec 03 19:49:23 crc kubenswrapper[4758]: I1203 19:49:23.914287 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-wdgqs_a1a68f23-f8d5-40c2-bca8-cf0f7ae19044/nmstate-operator/0.log" Dec 03 19:49:40 crc kubenswrapper[4758]: I1203 19:49:40.774389 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-vpzjk_01bb709f-7c93-45e7-b900-802665c0225f/kube-rbac-proxy/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.069453 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-frr-files/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.182485 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-vpzjk_01bb709f-7c93-45e7-b900-802665c0225f/controller/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.297274 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-reloader/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.325187 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-frr-files/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.374021 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-metrics/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.419162 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-reloader/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.606647 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-metrics/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.607729 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-reloader/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.622601 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-frr-files/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.683879 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-metrics/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.820811 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-frr-files/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.833441 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-metrics/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.876181 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/cp-reloader/0.log" Dec 03 19:49:41 crc kubenswrapper[4758]: I1203 19:49:41.940057 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/controller/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.104571 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/frr-metrics/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.123033 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/kube-rbac-proxy/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.181148 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/kube-rbac-proxy-frr/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.390374 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/reloader/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.413709 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-nqt6f_fa19d1a0-7212-4b46-8877-9e4e09f1f775/frr-k8s-webhook-server/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.669491 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-77cf4cbccf-5wfbn_3b2b2fc1-2f85-4c3b-956c-27378e0c4e5e/manager/0.log" Dec 03 19:49:42 crc kubenswrapper[4758]: I1203 19:49:42.852957 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-547d88fbfd-ns59r_9cfc8f20-5774-4fb3-ab06-179070a49c4c/webhook-server/0.log" Dec 03 19:49:43 crc kubenswrapper[4758]: I1203 19:49:42.978339 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zb4zw_56cd1dfc-f881-441e-b231-d2783ddcf0f9/kube-rbac-proxy/0.log" Dec 03 19:49:44 crc kubenswrapper[4758]: I1203 19:49:44.398842 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zb4zw_56cd1dfc-f881-441e-b231-d2783ddcf0f9/speaker/0.log" Dec 03 19:49:46 crc kubenswrapper[4758]: I1203 19:49:46.143646 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxqxh_272d2f0a-3217-4e67-82f3-43dc7aa334f5/frr/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.200824 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/util/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.443363 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/pull/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.473143 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/pull/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.473332 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/util/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.730456 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/util/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.737777 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/extract/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.762313 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apq8c8_24b483db-d9a3-4e07-ac9b-8f5c2b908bad/pull/0.log" Dec 03 19:50:00 crc kubenswrapper[4758]: I1203 19:50:00.951283 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/util/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.210359 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/util/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.228554 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/pull/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.229292 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/pull/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.428334 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/pull/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.440641 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/util/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.475921 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6fxp2_ad1568bc-9c16-4426-b740-c6ac65af84c0/extract/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.623282 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/util/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.852748 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/pull/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.891798 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/pull/0.log" Dec 03 19:50:01 crc kubenswrapper[4758]: I1203 19:50:01.899318 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/util/0.log" Dec 03 19:50:02 crc kubenswrapper[4758]: I1203 19:50:02.122371 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/util/0.log" Dec 03 19:50:02 crc kubenswrapper[4758]: I1203 19:50:02.164075 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/pull/0.log" Dec 03 19:50:02 crc kubenswrapper[4758]: I1203 19:50:02.173725 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwkk5_94b4bf6d-75a9-4b05-bb33-039174ec7a30/extract/0.log" Dec 03 19:50:02 crc kubenswrapper[4758]: I1203 19:50:02.839185 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/util/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.048865 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/util/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.102171 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/pull/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.102365 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/pull/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.334147 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/util/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.371471 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/extract/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.380451 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bq89h_71ef144e-1cc4-4f43-bf14-16237cf240b4/pull/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.565574 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/extract-utilities/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.776344 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/extract-content/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.810851 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/extract-utilities/0.log" Dec 03 19:50:03 crc kubenswrapper[4758]: I1203 19:50:03.822988 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/extract-content/0.log" Dec 03 19:50:04 crc kubenswrapper[4758]: I1203 19:50:04.134209 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/extract-content/0.log" Dec 03 19:50:04 crc kubenswrapper[4758]: I1203 19:50:04.134335 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/extract-utilities/0.log" Dec 03 19:50:04 crc kubenswrapper[4758]: I1203 19:50:04.247809 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/extract-utilities/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.103899 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/extract-content/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.106627 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/extract-utilities/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.147444 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/extract-content/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.415006 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/extract-content/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.433004 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/extract-utilities/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.678254 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fgbb5_6a4c5544-6114-40e4-8d89-f1a60bc1708a/marketplace-operator/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.779056 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/extract-utilities/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.912820 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-26n6q_3a247927-7fef-4538-b8bb-d3b2d146e318/registry-server/0.log" Dec 03 19:50:05 crc kubenswrapper[4758]: I1203 19:50:05.987602 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/extract-utilities/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.019398 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/extract-content/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.066059 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/extract-content/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.377947 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/extract-content/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.427219 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/extract-utilities/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.634118 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/extract-utilities/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.917752 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/extract-utilities/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.943045 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/extract-content/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.965638 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/extract-content/0.log" Dec 03 19:50:06 crc kubenswrapper[4758]: I1203 19:50:06.968304 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t52js_6d9c8c5c-e835-4f95-8c0d-7e6c7cd1f1a6/registry-server/0.log" Dec 03 19:50:07 crc kubenswrapper[4758]: I1203 19:50:07.176064 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/extract-utilities/0.log" Dec 03 19:50:07 crc kubenswrapper[4758]: I1203 19:50:07.233028 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/extract-content/0.log" Dec 03 19:50:07 crc kubenswrapper[4758]: I1203 19:50:07.725333 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-spff6_dc4501b8-839c-4cb5-a1a3-0b1a140e0c7f/registry-server/0.log" Dec 03 19:50:08 crc kubenswrapper[4758]: I1203 19:50:08.026150 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mmg7_aa3a0dc1-e4ad-41c2-b2c3-45c1831d4e24/registry-server/0.log" Dec 03 19:50:21 crc kubenswrapper[4758]: I1203 19:50:21.783086 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-rkkv8_373f7373-1c0a-4261-81b5-03a8efcf147e/prometheus-operator/0.log" Dec 03 19:50:22 crc kubenswrapper[4758]: I1203 19:50:22.186233 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6976796f4b-dkt4b_18a96e71-fd5d-4816-a001-92ae35d94c17/prometheus-operator-admission-webhook/0.log" Dec 03 19:50:22 crc kubenswrapper[4758]: I1203 19:50:22.241857 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6976796f4b-tgltn_aa3f69cb-5012-4879-864e-5826e71f2aa8/prometheus-operator-admission-webhook/0.log" Dec 03 19:50:22 crc kubenswrapper[4758]: I1203 19:50:22.473047 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-h79l8_424f6469-230e-4f6b-b257-7ad7feecd589/perses-operator/0.log" Dec 03 19:50:22 crc kubenswrapper[4758]: I1203 19:50:22.507241 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-r27q7_dd43e0c9-4231-459e-8d32-eddb92573c5b/operator/0.log" Dec 03 19:50:41 crc kubenswrapper[4758]: I1203 19:50:41.394330 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:50:41 crc kubenswrapper[4758]: I1203 19:50:41.394986 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:52.995453 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vphln"] Dec 03 19:50:53 crc kubenswrapper[4758]: E1203 19:50:52.996622 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="registry-server" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:52.996646 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="registry-server" Dec 03 19:50:53 crc kubenswrapper[4758]: E1203 19:50:52.996700 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="extract-content" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:52.996710 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="extract-content" Dec 03 19:50:53 crc kubenswrapper[4758]: E1203 19:50:52.996739 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="extract-utilities" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:52.996748 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="extract-utilities" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:52.997035 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed061115-c734-48b2-b201-2e88f72b1ede" containerName="registry-server" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.007522 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.046979 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vphln"] Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.140079 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-utilities\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.140295 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm5pm\" (UniqueName: \"kubernetes.io/projected/554799b3-b539-4f95-80ed-4807075a5095-kube-api-access-lm5pm\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.140474 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-catalog-content\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.242136 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-utilities\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.242383 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm5pm\" (UniqueName: \"kubernetes.io/projected/554799b3-b539-4f95-80ed-4807075a5095-kube-api-access-lm5pm\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.242670 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-catalog-content\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.242757 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-utilities\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.243432 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-catalog-content\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.270343 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm5pm\" (UniqueName: \"kubernetes.io/projected/554799b3-b539-4f95-80ed-4807075a5095-kube-api-access-lm5pm\") pod \"redhat-operators-vphln\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:53 crc kubenswrapper[4758]: I1203 19:50:53.359677 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:50:54 crc kubenswrapper[4758]: I1203 19:50:54.088163 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vphln"] Dec 03 19:50:54 crc kubenswrapper[4758]: I1203 19:50:54.840168 4758 generic.go:334] "Generic (PLEG): container finished" podID="554799b3-b539-4f95-80ed-4807075a5095" containerID="15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705" exitCode=0 Dec 03 19:50:54 crc kubenswrapper[4758]: I1203 19:50:54.840809 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerDied","Data":"15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705"} Dec 03 19:50:54 crc kubenswrapper[4758]: I1203 19:50:54.840842 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerStarted","Data":"1d4521fe86c27085a3fe37f4e41032587c26f362bde11bcc845538801bacc399"} Dec 03 19:50:54 crc kubenswrapper[4758]: I1203 19:50:54.848173 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:50:56 crc kubenswrapper[4758]: I1203 19:50:56.867079 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerStarted","Data":"886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d"} Dec 03 19:51:00 crc kubenswrapper[4758]: I1203 19:51:00.913573 4758 generic.go:334] "Generic (PLEG): container finished" podID="554799b3-b539-4f95-80ed-4807075a5095" containerID="886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d" exitCode=0 Dec 03 19:51:00 crc kubenswrapper[4758]: I1203 19:51:00.913724 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerDied","Data":"886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d"} Dec 03 19:51:02 crc kubenswrapper[4758]: I1203 19:51:02.936719 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerStarted","Data":"9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830"} Dec 03 19:51:02 crc kubenswrapper[4758]: I1203 19:51:02.965055 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vphln" podStartSLOduration=3.749243399 podStartE2EDuration="10.965032372s" podCreationTimestamp="2025-12-03 19:50:52 +0000 UTC" firstStartedPulling="2025-12-03 19:50:54.847863011 +0000 UTC m=+10510.049239872" lastFinishedPulling="2025-12-03 19:51:02.063651984 +0000 UTC m=+10517.265028845" observedRunningTime="2025-12-03 19:51:02.957896302 +0000 UTC m=+10518.159273163" watchObservedRunningTime="2025-12-03 19:51:02.965032372 +0000 UTC m=+10518.166409233" Dec 03 19:51:03 crc kubenswrapper[4758]: I1203 19:51:03.361143 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:51:03 crc kubenswrapper[4758]: I1203 19:51:03.361212 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:51:04 crc kubenswrapper[4758]: I1203 19:51:04.421049 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vphln" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="registry-server" probeResult="failure" output=< Dec 03 19:51:04 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Dec 03 19:51:04 crc kubenswrapper[4758]: > Dec 03 19:51:11 crc kubenswrapper[4758]: I1203 19:51:11.394708 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:51:11 crc kubenswrapper[4758]: I1203 19:51:11.395337 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:51:13 crc kubenswrapper[4758]: I1203 19:51:13.419495 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:51:13 crc kubenswrapper[4758]: I1203 19:51:13.479270 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:51:13 crc kubenswrapper[4758]: I1203 19:51:13.664751 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vphln"] Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.093556 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vphln" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="registry-server" containerID="cri-o://9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830" gracePeriod=2 Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.720387 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.871624 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm5pm\" (UniqueName: \"kubernetes.io/projected/554799b3-b539-4f95-80ed-4807075a5095-kube-api-access-lm5pm\") pod \"554799b3-b539-4f95-80ed-4807075a5095\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.872054 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-utilities\") pod \"554799b3-b539-4f95-80ed-4807075a5095\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.872086 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-catalog-content\") pod \"554799b3-b539-4f95-80ed-4807075a5095\" (UID: \"554799b3-b539-4f95-80ed-4807075a5095\") " Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.873584 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-utilities" (OuterVolumeSpecName: "utilities") pod "554799b3-b539-4f95-80ed-4807075a5095" (UID: "554799b3-b539-4f95-80ed-4807075a5095"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.881794 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/554799b3-b539-4f95-80ed-4807075a5095-kube-api-access-lm5pm" (OuterVolumeSpecName: "kube-api-access-lm5pm") pod "554799b3-b539-4f95-80ed-4807075a5095" (UID: "554799b3-b539-4f95-80ed-4807075a5095"). InnerVolumeSpecName "kube-api-access-lm5pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.975229 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm5pm\" (UniqueName: \"kubernetes.io/projected/554799b3-b539-4f95-80ed-4807075a5095-kube-api-access-lm5pm\") on node \"crc\" DevicePath \"\"" Dec 03 19:51:15 crc kubenswrapper[4758]: I1203 19:51:15.975269 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.006053 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "554799b3-b539-4f95-80ed-4807075a5095" (UID: "554799b3-b539-4f95-80ed-4807075a5095"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.077326 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554799b3-b539-4f95-80ed-4807075a5095-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.106530 4758 generic.go:334] "Generic (PLEG): container finished" podID="554799b3-b539-4f95-80ed-4807075a5095" containerID="9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830" exitCode=0 Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.106612 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerDied","Data":"9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830"} Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.106663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vphln" event={"ID":"554799b3-b539-4f95-80ed-4807075a5095","Type":"ContainerDied","Data":"1d4521fe86c27085a3fe37f4e41032587c26f362bde11bcc845538801bacc399"} Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.106714 4758 scope.go:117] "RemoveContainer" containerID="9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.106989 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vphln" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.134189 4758 scope.go:117] "RemoveContainer" containerID="886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.158668 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vphln"] Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.168579 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vphln"] Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.180932 4758 scope.go:117] "RemoveContainer" containerID="15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.224041 4758 scope.go:117] "RemoveContainer" containerID="9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830" Dec 03 19:51:16 crc kubenswrapper[4758]: E1203 19:51:16.224599 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830\": container with ID starting with 9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830 not found: ID does not exist" containerID="9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.224655 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830"} err="failed to get container status \"9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830\": rpc error: code = NotFound desc = could not find container \"9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830\": container with ID starting with 9f5f621109a82e142c10b560b9c3cb9d287c645c5e3b0d287f6e7800067ff830 not found: ID does not exist" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.224707 4758 scope.go:117] "RemoveContainer" containerID="886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d" Dec 03 19:51:16 crc kubenswrapper[4758]: E1203 19:51:16.226332 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d\": container with ID starting with 886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d not found: ID does not exist" containerID="886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.226376 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d"} err="failed to get container status \"886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d\": rpc error: code = NotFound desc = could not find container \"886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d\": container with ID starting with 886c015bf420434272df90aab1545c6934194a1a1650eddb16cff70eae3f457d not found: ID does not exist" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.226407 4758 scope.go:117] "RemoveContainer" containerID="15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705" Dec 03 19:51:16 crc kubenswrapper[4758]: E1203 19:51:16.226898 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705\": container with ID starting with 15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705 not found: ID does not exist" containerID="15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705" Dec 03 19:51:16 crc kubenswrapper[4758]: I1203 19:51:16.226928 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705"} err="failed to get container status \"15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705\": rpc error: code = NotFound desc = could not find container \"15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705\": container with ID starting with 15e3f10c075cc01cf5503ab0ea6985b8f66267134a37fa7d3a0baf694ea44705 not found: ID does not exist" Dec 03 19:51:17 crc kubenswrapper[4758]: I1203 19:51:17.127302 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="554799b3-b539-4f95-80ed-4807075a5095" path="/var/lib/kubelet/pods/554799b3-b539-4f95-80ed-4807075a5095/volumes" Dec 03 19:51:41 crc kubenswrapper[4758]: I1203 19:51:41.394743 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:51:41 crc kubenswrapper[4758]: I1203 19:51:41.395374 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:51:41 crc kubenswrapper[4758]: I1203 19:51:41.395429 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:51:41 crc kubenswrapper[4758]: I1203 19:51:41.396446 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e782b9e9687f9787128adb5b32b6b6158f798ff4567b2a8122c1b1aad4fc6275"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:51:41 crc kubenswrapper[4758]: I1203 19:51:41.396511 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://e782b9e9687f9787128adb5b32b6b6158f798ff4567b2a8122c1b1aad4fc6275" gracePeriod=600 Dec 03 19:51:42 crc kubenswrapper[4758]: I1203 19:51:42.439082 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="e782b9e9687f9787128adb5b32b6b6158f798ff4567b2a8122c1b1aad4fc6275" exitCode=0 Dec 03 19:51:42 crc kubenswrapper[4758]: I1203 19:51:42.439193 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"e782b9e9687f9787128adb5b32b6b6158f798ff4567b2a8122c1b1aad4fc6275"} Dec 03 19:51:42 crc kubenswrapper[4758]: I1203 19:51:42.439794 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerStarted","Data":"529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03"} Dec 03 19:51:42 crc kubenswrapper[4758]: I1203 19:51:42.439820 4758 scope.go:117] "RemoveContainer" containerID="1773615e8853db294f6e6e5cada1a5e652eb6d2577f53cdc4c38d17e512da727" Dec 03 19:52:58 crc kubenswrapper[4758]: I1203 19:52:58.350799 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb2432eb-1022-4475-ba37-c91e5435f875" containerID="367f6f1f74a93694c8e369ae18ec950d5c2e0c779556c72786652f642eb02cc4" exitCode=0 Dec 03 19:52:58 crc kubenswrapper[4758]: I1203 19:52:58.350885 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s29l4/must-gather-48wk6" event={"ID":"cb2432eb-1022-4475-ba37-c91e5435f875","Type":"ContainerDied","Data":"367f6f1f74a93694c8e369ae18ec950d5c2e0c779556c72786652f642eb02cc4"} Dec 03 19:52:58 crc kubenswrapper[4758]: I1203 19:52:58.351994 4758 scope.go:117] "RemoveContainer" containerID="367f6f1f74a93694c8e369ae18ec950d5c2e0c779556c72786652f642eb02cc4" Dec 03 19:52:59 crc kubenswrapper[4758]: I1203 19:52:59.403278 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s29l4_must-gather-48wk6_cb2432eb-1022-4475-ba37-c91e5435f875/gather/0.log" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.031547 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s29l4/must-gather-48wk6"] Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.032737 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-s29l4/must-gather-48wk6" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="copy" containerID="cri-o://133d6a89ec23246e89daa027b2760c2afadcf1e81c285f7346e3764a365a7d37" gracePeriod=2 Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.042751 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s29l4/must-gather-48wk6"] Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.468559 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s29l4_must-gather-48wk6_cb2432eb-1022-4475-ba37-c91e5435f875/copy/0.log" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.469329 4758 generic.go:334] "Generic (PLEG): container finished" podID="cb2432eb-1022-4475-ba37-c91e5435f875" containerID="133d6a89ec23246e89daa027b2760c2afadcf1e81c285f7346e3764a365a7d37" exitCode=143 Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.469387 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1277f4a6eaad6cb3ae2aefad9669275b4957537a38df2444fe5efa814882db5e" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.562370 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s29l4_must-gather-48wk6_cb2432eb-1022-4475-ba37-c91e5435f875/copy/0.log" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.562967 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.710233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl2zx\" (UniqueName: \"kubernetes.io/projected/cb2432eb-1022-4475-ba37-c91e5435f875-kube-api-access-nl2zx\") pod \"cb2432eb-1022-4475-ba37-c91e5435f875\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.710497 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cb2432eb-1022-4475-ba37-c91e5435f875-must-gather-output\") pod \"cb2432eb-1022-4475-ba37-c91e5435f875\" (UID: \"cb2432eb-1022-4475-ba37-c91e5435f875\") " Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.716453 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb2432eb-1022-4475-ba37-c91e5435f875-kube-api-access-nl2zx" (OuterVolumeSpecName: "kube-api-access-nl2zx") pod "cb2432eb-1022-4475-ba37-c91e5435f875" (UID: "cb2432eb-1022-4475-ba37-c91e5435f875"). InnerVolumeSpecName "kube-api-access-nl2zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.813183 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl2zx\" (UniqueName: \"kubernetes.io/projected/cb2432eb-1022-4475-ba37-c91e5435f875-kube-api-access-nl2zx\") on node \"crc\" DevicePath \"\"" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.910904 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb2432eb-1022-4475-ba37-c91e5435f875-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cb2432eb-1022-4475-ba37-c91e5435f875" (UID: "cb2432eb-1022-4475-ba37-c91e5435f875"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:53:08 crc kubenswrapper[4758]: I1203 19:53:08.917534 4758 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cb2432eb-1022-4475-ba37-c91e5435f875-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 19:53:09 crc kubenswrapper[4758]: I1203 19:53:09.144981 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" path="/var/lib/kubelet/pods/cb2432eb-1022-4475-ba37-c91e5435f875/volumes" Dec 03 19:53:09 crc kubenswrapper[4758]: E1203 19:53:09.236817 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb2432eb_1022_4475_ba37_c91e5435f875.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:53:09 crc kubenswrapper[4758]: I1203 19:53:09.478959 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s29l4/must-gather-48wk6" Dec 03 19:53:23 crc kubenswrapper[4758]: I1203 19:53:23.164221 4758 scope.go:117] "RemoveContainer" containerID="367f6f1f74a93694c8e369ae18ec950d5c2e0c779556c72786652f642eb02cc4" Dec 03 19:53:23 crc kubenswrapper[4758]: I1203 19:53:23.267312 4758 scope.go:117] "RemoveContainer" containerID="133d6a89ec23246e89daa027b2760c2afadcf1e81c285f7346e3764a365a7d37" Dec 03 19:53:41 crc kubenswrapper[4758]: I1203 19:53:41.394880 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:53:41 crc kubenswrapper[4758]: I1203 19:53:41.395513 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:54:11 crc kubenswrapper[4758]: I1203 19:54:11.395179 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:54:11 crc kubenswrapper[4758]: I1203 19:54:11.395650 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.394157 4758 patch_prober.go:28] interesting pod/machine-config-daemon-fdssg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.395281 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.395326 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.396179 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03"} pod="openshift-machine-config-operator/machine-config-daemon-fdssg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.396233 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerName="machine-config-daemon" containerID="cri-o://529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" gracePeriod=600 Dec 03 19:54:41 crc kubenswrapper[4758]: E1203 19:54:41.528310 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.569488 4758 generic.go:334] "Generic (PLEG): container finished" podID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" exitCode=0 Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.569539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" event={"ID":"2e7d0dba-00e7-4a9d-915e-e31788c7a2bb","Type":"ContainerDied","Data":"529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03"} Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.569572 4758 scope.go:117] "RemoveContainer" containerID="e782b9e9687f9787128adb5b32b6b6158f798ff4567b2a8122c1b1aad4fc6275" Dec 03 19:54:41 crc kubenswrapper[4758]: I1203 19:54:41.570401 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:54:41 crc kubenswrapper[4758]: E1203 19:54:41.570708 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:54:53 crc kubenswrapper[4758]: I1203 19:54:53.114661 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:54:53 crc kubenswrapper[4758]: E1203 19:54:53.115442 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:55:07 crc kubenswrapper[4758]: I1203 19:55:07.114194 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:55:07 crc kubenswrapper[4758]: E1203 19:55:07.115065 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.492756 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vxkj5"] Dec 03 19:55:15 crc kubenswrapper[4758]: E1203 19:55:15.493996 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="copy" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494014 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="copy" Dec 03 19:55:15 crc kubenswrapper[4758]: E1203 19:55:15.494054 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="extract-content" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494063 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="extract-content" Dec 03 19:55:15 crc kubenswrapper[4758]: E1203 19:55:15.494079 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="extract-utilities" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494089 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="extract-utilities" Dec 03 19:55:15 crc kubenswrapper[4758]: E1203 19:55:15.494097 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="gather" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494105 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="gather" Dec 03 19:55:15 crc kubenswrapper[4758]: E1203 19:55:15.494128 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="registry-server" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494135 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="registry-server" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494395 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="copy" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494415 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="554799b3-b539-4f95-80ed-4807075a5095" containerName="registry-server" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.494429 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb2432eb-1022-4475-ba37-c91e5435f875" containerName="gather" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.496355 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.512836 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vxkj5"] Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.605394 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-utilities\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.605526 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-catalog-content\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.605567 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pcdv\" (UniqueName: \"kubernetes.io/projected/4e012f79-7b65-4990-8e18-579d63bdd147-kube-api-access-4pcdv\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.708205 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-utilities\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.708329 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-catalog-content\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.708368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pcdv\" (UniqueName: \"kubernetes.io/projected/4e012f79-7b65-4990-8e18-579d63bdd147-kube-api-access-4pcdv\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.709345 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-utilities\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.709634 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-catalog-content\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.738709 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pcdv\" (UniqueName: \"kubernetes.io/projected/4e012f79-7b65-4990-8e18-579d63bdd147-kube-api-access-4pcdv\") pod \"certified-operators-vxkj5\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:15 crc kubenswrapper[4758]: I1203 19:55:15.835502 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:16 crc kubenswrapper[4758]: I1203 19:55:16.586927 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vxkj5"] Dec 03 19:55:16 crc kubenswrapper[4758]: I1203 19:55:16.987355 4758 generic.go:334] "Generic (PLEG): container finished" podID="4e012f79-7b65-4990-8e18-579d63bdd147" containerID="b1570a1649ba39c74402a5a8c77374527b8718f24fadd34c833fcb8a6fb33b0a" exitCode=0 Dec 03 19:55:16 crc kubenswrapper[4758]: I1203 19:55:16.987432 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerDied","Data":"b1570a1649ba39c74402a5a8c77374527b8718f24fadd34c833fcb8a6fb33b0a"} Dec 03 19:55:16 crc kubenswrapper[4758]: I1203 19:55:16.987909 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerStarted","Data":"249b476f6a87d93862eb5916010615cf94b7b8cdca99815db7d72b944acfe6a4"} Dec 03 19:55:18 crc kubenswrapper[4758]: I1203 19:55:18.008132 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerStarted","Data":"15c2c66ec93dd6c3f192d3f413efb72eb519aeda82ee433c851cbe9ba2d5b8c7"} Dec 03 19:55:19 crc kubenswrapper[4758]: I1203 19:55:19.023098 4758 generic.go:334] "Generic (PLEG): container finished" podID="4e012f79-7b65-4990-8e18-579d63bdd147" containerID="15c2c66ec93dd6c3f192d3f413efb72eb519aeda82ee433c851cbe9ba2d5b8c7" exitCode=0 Dec 03 19:55:19 crc kubenswrapper[4758]: I1203 19:55:19.023160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerDied","Data":"15c2c66ec93dd6c3f192d3f413efb72eb519aeda82ee433c851cbe9ba2d5b8c7"} Dec 03 19:55:20 crc kubenswrapper[4758]: I1203 19:55:20.045910 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerStarted","Data":"49e54566ca5c539f2305b7ebe2a30d45a7ce4833705e8ca0a635294ffcdeaa25"} Dec 03 19:55:20 crc kubenswrapper[4758]: I1203 19:55:20.074290 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vxkj5" podStartSLOduration=2.639669074 podStartE2EDuration="5.074270058s" podCreationTimestamp="2025-12-03 19:55:15 +0000 UTC" firstStartedPulling="2025-12-03 19:55:16.989555539 +0000 UTC m=+10772.190932400" lastFinishedPulling="2025-12-03 19:55:19.424156523 +0000 UTC m=+10774.625533384" observedRunningTime="2025-12-03 19:55:20.062252587 +0000 UTC m=+10775.263629468" watchObservedRunningTime="2025-12-03 19:55:20.074270058 +0000 UTC m=+10775.275646919" Dec 03 19:55:21 crc kubenswrapper[4758]: I1203 19:55:21.114129 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:55:21 crc kubenswrapper[4758]: E1203 19:55:21.114377 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:55:25 crc kubenswrapper[4758]: I1203 19:55:25.836778 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:25 crc kubenswrapper[4758]: I1203 19:55:25.837421 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:25 crc kubenswrapper[4758]: I1203 19:55:25.889748 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:26 crc kubenswrapper[4758]: I1203 19:55:26.156035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:26 crc kubenswrapper[4758]: I1203 19:55:26.208969 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vxkj5"] Dec 03 19:55:28 crc kubenswrapper[4758]: I1203 19:55:28.129542 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vxkj5" podUID="4e012f79-7b65-4990-8e18-579d63bdd147" containerName="registry-server" containerID="cri-o://49e54566ca5c539f2305b7ebe2a30d45a7ce4833705e8ca0a635294ffcdeaa25" gracePeriod=2 Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.152168 4758 generic.go:334] "Generic (PLEG): container finished" podID="4e012f79-7b65-4990-8e18-579d63bdd147" containerID="49e54566ca5c539f2305b7ebe2a30d45a7ce4833705e8ca0a635294ffcdeaa25" exitCode=0 Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.152418 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerDied","Data":"49e54566ca5c539f2305b7ebe2a30d45a7ce4833705e8ca0a635294ffcdeaa25"} Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.154118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxkj5" event={"ID":"4e012f79-7b65-4990-8e18-579d63bdd147","Type":"ContainerDied","Data":"249b476f6a87d93862eb5916010615cf94b7b8cdca99815db7d72b944acfe6a4"} Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.154237 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="249b476f6a87d93862eb5916010615cf94b7b8cdca99815db7d72b944acfe6a4" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.198849 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.287173 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-utilities\") pod \"4e012f79-7b65-4990-8e18-579d63bdd147\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.287563 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-catalog-content\") pod \"4e012f79-7b65-4990-8e18-579d63bdd147\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.287852 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pcdv\" (UniqueName: \"kubernetes.io/projected/4e012f79-7b65-4990-8e18-579d63bdd147-kube-api-access-4pcdv\") pod \"4e012f79-7b65-4990-8e18-579d63bdd147\" (UID: \"4e012f79-7b65-4990-8e18-579d63bdd147\") " Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.290070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-utilities" (OuterVolumeSpecName: "utilities") pod "4e012f79-7b65-4990-8e18-579d63bdd147" (UID: "4e012f79-7b65-4990-8e18-579d63bdd147"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.298548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e012f79-7b65-4990-8e18-579d63bdd147-kube-api-access-4pcdv" (OuterVolumeSpecName: "kube-api-access-4pcdv") pod "4e012f79-7b65-4990-8e18-579d63bdd147" (UID: "4e012f79-7b65-4990-8e18-579d63bdd147"). InnerVolumeSpecName "kube-api-access-4pcdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.346055 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e012f79-7b65-4990-8e18-579d63bdd147" (UID: "4e012f79-7b65-4990-8e18-579d63bdd147"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.391383 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.391430 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pcdv\" (UniqueName: \"kubernetes.io/projected/4e012f79-7b65-4990-8e18-579d63bdd147-kube-api-access-4pcdv\") on node \"crc\" DevicePath \"\"" Dec 03 19:55:29 crc kubenswrapper[4758]: I1203 19:55:29.391440 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e012f79-7b65-4990-8e18-579d63bdd147-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:55:30 crc kubenswrapper[4758]: I1203 19:55:30.162990 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxkj5" Dec 03 19:55:30 crc kubenswrapper[4758]: I1203 19:55:30.204894 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vxkj5"] Dec 03 19:55:30 crc kubenswrapper[4758]: I1203 19:55:30.215148 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vxkj5"] Dec 03 19:55:31 crc kubenswrapper[4758]: I1203 19:55:31.126854 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e012f79-7b65-4990-8e18-579d63bdd147" path="/var/lib/kubelet/pods/4e012f79-7b65-4990-8e18-579d63bdd147/volumes" Dec 03 19:55:33 crc kubenswrapper[4758]: I1203 19:55:33.114982 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:55:33 crc kubenswrapper[4758]: E1203 19:55:33.115604 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:55:46 crc kubenswrapper[4758]: I1203 19:55:46.115028 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:55:46 crc kubenswrapper[4758]: E1203 19:55:46.117191 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" Dec 03 19:56:00 crc kubenswrapper[4758]: I1203 19:56:00.114302 4758 scope.go:117] "RemoveContainer" containerID="529ebdf17c48e4f9cb8c684d09f3013c072dfbc628431249026e435589c1cb03" Dec 03 19:56:00 crc kubenswrapper[4758]: E1203 19:56:00.115247 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fdssg_openshift-machine-config-operator(2e7d0dba-00e7-4a9d-915e-e31788c7a2bb)\"" pod="openshift-machine-config-operator/machine-config-daemon-fdssg" podUID="2e7d0dba-00e7-4a9d-915e-e31788c7a2bb" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114112740024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114112741017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114065226016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114065226015457 5ustar corecore